2026-03-06T22:12:23.410 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-06T22:12:23.414 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-06T22:12:23.441 INFO:teuthology.run:Config: archive_path: /archive/irq0-2026-03-06_20:21:59-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/382 branch: cobaltcore-storage-v19.2.3-fasttrack-5 description: orch:cephadm:smoke-roleless/{0-distro/ubuntu_22.04 1-start 2-services/nfs-haproxy-proto 3-final} email: null first_in_suite: false flavor: default job_id: '382' ktype: distro last_in_suite: false machine_type: vps name: irq0-2026-03-06_20:21:59-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps no_nested_subset: false openstack: - volumes: count: 4 size: 10 os_type: ubuntu os_version: '22.04' overrides: admin_socket: branch: cobaltcore-storage-v19.2.3-fasttrack-5 ansible.cephlab: branch: main repo: https://github.com/kshtsk/ceph-cm-ansible.git skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: Europe/Berlin ceph: conf: mgr: debug mgr: 20 debug ms: 1 mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 osd shutdown pgref assert: true flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - CEPHADM_DAEMON_PLACE_FAIL - CEPHADM_FAILED_DAEMON log-only-match: - CEPHADM_ sha1: 340d3c24fc6ae7529322dc7ccee6c6cb2589da0a ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} cephadm: cephadm_binary_url: https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm containers: image: harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 install: ceph: flavor: default sha1: 340d3c24fc6ae7529322dc7ccee6c6cb2589da0a extra_system_packages: deb: - python3-xmltodict - s3cmd rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - s3cmd repos: - name: ceph-source priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-39-g340d3c24fc6/el9.clyso/SRPMS - name: ceph-noarch priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-39-g340d3c24fc6/el9.clyso/noarch - name: ceph priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-39-g340d3c24fc6/el9.clyso/x86_64 workunit: branch: tt-19.2.3-fasttrack-5-no-nvme-loop sha1: b952d7263a165ada4530724b87fab57a8f3f547b owner: irq0 priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - client.0 - - host.b - client.1 seed: 9421 sha1: 340d3c24fc6ae7529322dc7ccee6c6cb2589da0a sleep_before_teardown: 0 suite: orch:cephadm:smoke-roleless suite_branch: tt-19.2.3-fasttrack-5-no-nvme-loop suite_path: /home/teuthos/src/github.com_kshtsk_ceph_b952d7263a165ada4530724b87fab57a8f3f547b/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: b952d7263a165ada4530724b87fab57a8f3f547b targets: vm03.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLQ1RSclccXlg9yxkwruHAoRUSUtCiOPDBGAUa6Z+P828llIB1cHol7eWThe7QkNMCJraOKl1CB45TB+Izbc5Y0= vm08.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBII9nclopw/qm3BB3bNzWyXX30fXTOAkRoOX7sZdqaZr2jx1r+JkOAp1tFRfvA13s+GiCt47VKODjS92NMa9ZIY= tasks: - cephadm: roleless: true - cephadm.shell: host.a: - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - vip: null - cephadm.shell: host.a: - ceph orch device ls --refresh - vip.exec: all-hosts: - systemctl stop nfs-server - cephadm.shell: host.a: - ceph fs volume create fs1 - ceph nfs cluster create happy --ingress --virtual-ip={{VIP0}} --ingress-mode=haproxy-protocol - ceph nfs export create cephfs --fsname fs1 --cluster-id happy --pseudo-path /d1 - cephadm.wait_for_service: service: nfs.happy - cephadm.wait_for_service: service: ingress.nfs.happy - vip.exec: host.a: - mkdir /mnt/happy - sleep 1 - mount -t nfs {{VIP0}}:/d1 /mnt/happy - echo test > /mnt/happy/testfile - sync - cephadm.shell: host.a: - stat -c '%u %g' /var/log/ceph | grep '167 167' - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - ceph orch ls | grep '^osd.all-available-devices ' teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-06_20:21:59 tube: vps user: irq0 verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.43333 2026-03-06T22:12:23.441 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_b952d7263a165ada4530724b87fab57a8f3f547b/qa; will attempt to use it 2026-03-06T22:12:23.442 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_b952d7263a165ada4530724b87fab57a8f3f547b/qa/tasks 2026-03-06T22:12:23.442 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-06T22:12:23.442 INFO:teuthology.task.internal:Saving configuration 2026-03-06T22:12:23.447 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-06T22:12:23.448 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-06T22:12:23.456 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm03.local', 'description': '/archive/irq0-2026-03-06_20:21:59-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/382', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-06 21:11:20.138952', 'locked_by': 'irq0', 'mac_address': '52:55:00:00:00:03', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLQ1RSclccXlg9yxkwruHAoRUSUtCiOPDBGAUa6Z+P828llIB1cHol7eWThe7QkNMCJraOKl1CB45TB+Izbc5Y0='} 2026-03-06T22:12:23.463 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm08.local', 'description': '/archive/irq0-2026-03-06_20:21:59-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/382', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-06 21:11:20.139421', 'locked_by': 'irq0', 'mac_address': '52:55:00:00:00:08', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBII9nclopw/qm3BB3bNzWyXX30fXTOAkRoOX7sZdqaZr2jx1r+JkOAp1tFRfvA13s+GiCt47VKODjS92NMa9ZIY='} 2026-03-06T22:12:23.463 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-06T22:12:23.464 INFO:teuthology.task.internal:roles: ubuntu@vm03.local - ['host.a', 'client.0'] 2026-03-06T22:12:23.464 INFO:teuthology.task.internal:roles: ubuntu@vm08.local - ['host.b', 'client.1'] 2026-03-06T22:12:23.464 INFO:teuthology.run_tasks:Running task console_log... 2026-03-06T22:12:23.472 DEBUG:teuthology.task.console_log:vm03 does not support IPMI; excluding 2026-03-06T22:12:23.479 DEBUG:teuthology.task.console_log:vm08 does not support IPMI; excluding 2026-03-06T22:12:23.479 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7fa89b5c7d90>, signals=[15]) 2026-03-06T22:12:23.479 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-06T22:12:23.480 INFO:teuthology.task.internal:Opening connections... 2026-03-06T22:12:23.480 DEBUG:teuthology.task.internal:connecting to ubuntu@vm03.local 2026-03-06T22:12:23.481 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm03.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T22:12:23.539 DEBUG:teuthology.task.internal:connecting to ubuntu@vm08.local 2026-03-06T22:12:23.539 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T22:12:23.600 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-06T22:12:23.601 DEBUG:teuthology.orchestra.run.vm03:> uname -m 2026-03-06T22:12:23.621 INFO:teuthology.orchestra.run.vm03.stdout:x86_64 2026-03-06T22:12:23.621 DEBUG:teuthology.orchestra.run.vm03:> cat /etc/os-release 2026-03-06T22:12:23.667 INFO:teuthology.orchestra.run.vm03.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-06T22:12:23.667 INFO:teuthology.orchestra.run.vm03.stdout:NAME="Ubuntu" 2026-03-06T22:12:23.667 INFO:teuthology.orchestra.run.vm03.stdout:VERSION_ID="22.04" 2026-03-06T22:12:23.667 INFO:teuthology.orchestra.run.vm03.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-06T22:12:23.667 INFO:teuthology.orchestra.run.vm03.stdout:VERSION_CODENAME=jammy 2026-03-06T22:12:23.667 INFO:teuthology.orchestra.run.vm03.stdout:ID=ubuntu 2026-03-06T22:12:23.667 INFO:teuthology.orchestra.run.vm03.stdout:ID_LIKE=debian 2026-03-06T22:12:23.667 INFO:teuthology.orchestra.run.vm03.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-06T22:12:23.667 INFO:teuthology.orchestra.run.vm03.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-06T22:12:23.667 INFO:teuthology.orchestra.run.vm03.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-06T22:12:23.667 INFO:teuthology.orchestra.run.vm03.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-06T22:12:23.667 INFO:teuthology.orchestra.run.vm03.stdout:UBUNTU_CODENAME=jammy 2026-03-06T22:12:23.667 INFO:teuthology.lock.ops:Updating vm03.local on lock server 2026-03-06T22:12:23.671 DEBUG:teuthology.orchestra.run.vm08:> uname -m 2026-03-06T22:12:23.674 INFO:teuthology.orchestra.run.vm08.stdout:x86_64 2026-03-06T22:12:23.674 DEBUG:teuthology.orchestra.run.vm08:> cat /etc/os-release 2026-03-06T22:12:23.720 INFO:teuthology.orchestra.run.vm08.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-06T22:12:23.720 INFO:teuthology.orchestra.run.vm08.stdout:NAME="Ubuntu" 2026-03-06T22:12:23.720 INFO:teuthology.orchestra.run.vm08.stdout:VERSION_ID="22.04" 2026-03-06T22:12:23.720 INFO:teuthology.orchestra.run.vm08.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-06T22:12:23.720 INFO:teuthology.orchestra.run.vm08.stdout:VERSION_CODENAME=jammy 2026-03-06T22:12:23.720 INFO:teuthology.orchestra.run.vm08.stdout:ID=ubuntu 2026-03-06T22:12:23.720 INFO:teuthology.orchestra.run.vm08.stdout:ID_LIKE=debian 2026-03-06T22:12:23.720 INFO:teuthology.orchestra.run.vm08.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-06T22:12:23.720 INFO:teuthology.orchestra.run.vm08.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-06T22:12:23.721 INFO:teuthology.orchestra.run.vm08.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-06T22:12:23.721 INFO:teuthology.orchestra.run.vm08.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-06T22:12:23.721 INFO:teuthology.orchestra.run.vm08.stdout:UBUNTU_CODENAME=jammy 2026-03-06T22:12:23.721 INFO:teuthology.lock.ops:Updating vm08.local on lock server 2026-03-06T22:12:23.726 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-06T22:12:23.728 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-06T22:12:23.728 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-06T22:12:23.728 DEBUG:teuthology.orchestra.run.vm03:> test '!' -e /home/ubuntu/cephtest 2026-03-06T22:12:23.730 DEBUG:teuthology.orchestra.run.vm08:> test '!' -e /home/ubuntu/cephtest 2026-03-06T22:12:23.764 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-06T22:12:23.765 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-06T22:12:23.765 DEBUG:teuthology.orchestra.run.vm03:> test -z $(ls -A /var/lib/ceph) 2026-03-06T22:12:23.773 DEBUG:teuthology.orchestra.run.vm08:> test -z $(ls -A /var/lib/ceph) 2026-03-06T22:12:23.775 INFO:teuthology.orchestra.run.vm03.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-06T22:12:23.809 INFO:teuthology.orchestra.run.vm08.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-06T22:12:23.809 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-06T22:12:23.818 DEBUG:teuthology.orchestra.run.vm03:> test -e /ceph-qa-ready 2026-03-06T22:12:23.821 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T22:12:24.254 DEBUG:teuthology.orchestra.run.vm08:> test -e /ceph-qa-ready 2026-03-06T22:12:24.258 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T22:12:24.504 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-06T22:12:24.505 INFO:teuthology.task.internal:Creating test directory... 2026-03-06T22:12:24.506 DEBUG:teuthology.orchestra.run.vm03:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-06T22:12:24.507 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-06T22:12:24.511 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-06T22:12:24.512 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-06T22:12:24.514 INFO:teuthology.task.internal:Creating archive directory... 2026-03-06T22:12:24.514 DEBUG:teuthology.orchestra.run.vm03:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-06T22:12:24.553 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-06T22:12:24.558 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-06T22:12:24.560 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-06T22:12:24.560 DEBUG:teuthology.orchestra.run.vm03:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-06T22:12:24.598 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T22:12:24.598 DEBUG:teuthology.orchestra.run.vm08:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-06T22:12:24.603 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T22:12:24.604 DEBUG:teuthology.orchestra.run.vm03:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-06T22:12:24.641 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-06T22:12:24.648 INFO:teuthology.orchestra.run.vm03.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T22:12:24.653 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T22:12:24.654 INFO:teuthology.orchestra.run.vm03.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T22:12:24.659 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T22:12:24.660 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-06T22:12:24.661 INFO:teuthology.task.internal:Configuring sudo... 2026-03-06T22:12:24.662 DEBUG:teuthology.orchestra.run.vm03:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-06T22:12:24.696 DEBUG:teuthology.orchestra.run.vm08:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-06T22:12:24.709 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-06T22:12:24.711 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-06T22:12:24.711 DEBUG:teuthology.orchestra.run.vm03:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-06T22:12:24.748 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-06T22:12:24.756 DEBUG:teuthology.orchestra.run.vm03:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T22:12:24.794 DEBUG:teuthology.orchestra.run.vm03:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T22:12:24.838 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-06T22:12:24.838 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-06T22:12:24.888 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T22:12:24.892 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T22:12:24.936 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T22:12:24.936 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-06T22:12:24.984 DEBUG:teuthology.orchestra.run.vm03:> sudo service rsyslog restart 2026-03-06T22:12:24.986 DEBUG:teuthology.orchestra.run.vm08:> sudo service rsyslog restart 2026-03-06T22:12:25.044 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-06T22:12:25.045 INFO:teuthology.task.internal:Starting timer... 2026-03-06T22:12:25.045 INFO:teuthology.run_tasks:Running task pcp... 2026-03-06T22:12:25.048 INFO:teuthology.run_tasks:Running task selinux... 2026-03-06T22:12:25.050 INFO:teuthology.task.selinux:Excluding vm03: VMs are not yet supported 2026-03-06T22:12:25.050 INFO:teuthology.task.selinux:Excluding vm08: VMs are not yet supported 2026-03-06T22:12:25.050 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-06T22:12:25.050 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-06T22:12:25.050 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-06T22:12:25.050 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-06T22:12:25.051 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'repo': 'https://github.com/kshtsk/ceph-cm-ansible.git', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'Europe/Berlin'}} 2026-03-06T22:12:25.052 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/kshtsk/ceph-cm-ansible.git 2026-03-06T22:12:25.053 INFO:teuthology.repo_utils:Fetching github.com_kshtsk_ceph-cm-ansible_main from origin 2026-03-06T22:12:25.630 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main to origin/main 2026-03-06T22:12:25.636 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-06T22:12:25.636 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "Europe/Berlin"}' -i /tmp/teuth_ansible_inventorygf5i1cq6 --limit vm03.local,vm08.local /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-06T22:14:49.175 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm03.local'), Remote(name='ubuntu@vm08.local')] 2026-03-06T22:14:49.175 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm03.local' 2026-03-06T22:14:49.176 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm03.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T22:14:49.242 DEBUG:teuthology.orchestra.run.vm03:> true 2026-03-06T22:14:49.476 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm03.local' 2026-03-06T22:14:49.485 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm08.local' 2026-03-06T22:14:49.485 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T22:14:49.549 DEBUG:teuthology.orchestra.run.vm08:> true 2026-03-06T22:14:49.808 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm08.local' 2026-03-06T22:14:49.808 INFO:teuthology.run_tasks:Running task clock... 2026-03-06T22:14:49.811 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-06T22:14:49.811 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-06T22:14:49.811 DEBUG:teuthology.orchestra.run.vm03:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T22:14:49.813 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-06T22:14:49.813 DEBUG:teuthology.orchestra.run.vm08:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T22:14:49.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-06T22:14:49.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: Command line: ntpd -gq 2026-03-06T22:14:49.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: ---------------------------------------------------- 2026-03-06T22:14:49.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: ntp-4 is maintained by Network Time Foundation, 2026-03-06T22:14:49.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-06T22:14:49.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: corporation. Support and training for ntp-4 are 2026-03-06T22:14:49.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: available at https://www.nwtime.org/support 2026-03-06T22:14:49.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: ---------------------------------------------------- 2026-03-06T22:14:49.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: proto: precision = 0.040 usec (-24) 2026-03-06T22:14:49.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: basedate set to 2022-02-04 2026-03-06T22:14:49.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: gps base set to 2022-02-06 (week 2196) 2026-03-06T22:14:49.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-06T22:14:49.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-06T22:14:49.831 INFO:teuthology.orchestra.run.vm03.stderr: 6 Mar 22:14:49 ntpd[15662]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 69 days ago 2026-03-06T22:14:49.831 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: Listen and drop on 0 v6wildcard [::]:123 2026-03-06T22:14:49.831 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-06T22:14:49.831 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: Listen normally on 2 lo 127.0.0.1:123 2026-03-06T22:14:49.831 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: Listen normally on 3 ens3 192.168.123.103:123 2026-03-06T22:14:49.831 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: Listen normally on 4 lo [::1]:123 2026-03-06T22:14:49.831 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:3%2]:123 2026-03-06T22:14:49.831 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:49 ntpd[15662]: Listening on routing socket on fd #22 for interface updates 2026-03-06T22:14:49.867 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-06T22:14:49.867 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: Command line: ntpd -gq 2026-03-06T22:14:49.867 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: ---------------------------------------------------- 2026-03-06T22:14:49.867 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: ntp-4 is maintained by Network Time Foundation, 2026-03-06T22:14:49.867 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-06T22:14:49.867 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: corporation. Support and training for ntp-4 are 2026-03-06T22:14:49.867 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: available at https://www.nwtime.org/support 2026-03-06T22:14:49.867 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: ---------------------------------------------------- 2026-03-06T22:14:49.868 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: proto: precision = 0.040 usec (-24) 2026-03-06T22:14:49.868 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: basedate set to 2022-02-04 2026-03-06T22:14:49.868 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: gps base set to 2022-02-06 (week 2196) 2026-03-06T22:14:49.868 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-06T22:14:49.868 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-06T22:14:49.868 INFO:teuthology.orchestra.run.vm08.stderr: 6 Mar 22:14:49 ntpd[15635]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 69 days ago 2026-03-06T22:14:49.869 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: Listen and drop on 0 v6wildcard [::]:123 2026-03-06T22:14:49.869 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-06T22:14:49.870 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: Listen normally on 2 lo 127.0.0.1:123 2026-03-06T22:14:49.870 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: Listen normally on 3 ens3 192.168.123.108:123 2026-03-06T22:14:49.870 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: Listen normally on 4 lo [::1]:123 2026-03-06T22:14:49.870 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:8%2]:123 2026-03-06T22:14:49.870 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:49 ntpd[15635]: Listening on routing socket on fd #22 for interface updates 2026-03-06T22:14:50.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:50 ntpd[15662]: Soliciting pool server 141.144.241.16 2026-03-06T22:14:50.868 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:50 ntpd[15635]: Soliciting pool server 141.144.241.16 2026-03-06T22:14:51.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:51 ntpd[15662]: Soliciting pool server 129.250.35.251 2026-03-06T22:14:51.868 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:51 ntpd[15635]: Soliciting pool server 129.250.35.251 2026-03-06T22:14:52.829 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:52 ntpd[15662]: Soliciting pool server 195.201.125.53 2026-03-06T22:14:52.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:52 ntpd[15662]: Soliciting pool server 172.236.195.26 2026-03-06T22:14:52.867 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:52 ntpd[15635]: Soliciting pool server 195.201.125.53 2026-03-06T22:14:52.867 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:52 ntpd[15635]: Soliciting pool server 172.236.195.26 2026-03-06T22:14:53.829 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:53 ntpd[15662]: Soliciting pool server 217.144.138.234 2026-03-06T22:14:53.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:53 ntpd[15662]: Soliciting pool server 139.162.156.95 2026-03-06T22:14:53.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:53 ntpd[15662]: Soliciting pool server 46.224.156.215 2026-03-06T22:14:53.866 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:53 ntpd[15635]: Soliciting pool server 217.144.138.234 2026-03-06T22:14:53.866 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:53 ntpd[15635]: Soliciting pool server 139.162.156.95 2026-03-06T22:14:53.866 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:53 ntpd[15635]: Soliciting pool server 46.224.156.215 2026-03-06T22:14:54.829 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:54 ntpd[15662]: Soliciting pool server 213.209.109.45 2026-03-06T22:14:54.829 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:54 ntpd[15662]: Soliciting pool server 79.133.44.142 2026-03-06T22:14:54.829 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:54 ntpd[15662]: Soliciting pool server 85.214.133.14 2026-03-06T22:14:54.830 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:54 ntpd[15662]: Soliciting pool server 91.189.91.157 2026-03-06T22:14:54.866 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:54 ntpd[15635]: Soliciting pool server 213.209.109.45 2026-03-06T22:14:54.866 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:54 ntpd[15635]: Soliciting pool server 79.133.44.142 2026-03-06T22:14:54.866 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:54 ntpd[15635]: Soliciting pool server 85.214.133.14 2026-03-06T22:14:54.866 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:54 ntpd[15635]: Soliciting pool server 91.189.91.157 2026-03-06T22:14:55.829 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:55 ntpd[15662]: Soliciting pool server 185.125.190.56 2026-03-06T22:14:55.829 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:55 ntpd[15662]: Soliciting pool server 82.165.178.31 2026-03-06T22:14:55.829 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:55 ntpd[15662]: Soliciting pool server 78.46.87.46 2026-03-06T22:14:55.865 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:55 ntpd[15635]: Soliciting pool server 185.125.190.56 2026-03-06T22:14:55.866 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:55 ntpd[15635]: Soliciting pool server 82.165.178.31 2026-03-06T22:14:55.866 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:55 ntpd[15635]: Soliciting pool server 78.46.87.46 2026-03-06T22:14:56.829 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:56 ntpd[15662]: Soliciting pool server 185.125.190.58 2026-03-06T22:14:56.829 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:56 ntpd[15662]: Soliciting pool server 85.215.189.120 2026-03-06T22:14:56.829 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:56 ntpd[15662]: Soliciting pool server 2603:c020:8017:3eff::123 2026-03-06T22:14:56.865 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:56 ntpd[15635]: Soliciting pool server 185.125.190.58 2026-03-06T22:14:56.865 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:56 ntpd[15635]: Soliciting pool server 85.215.189.120 2026-03-06T22:14:56.865 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:56 ntpd[15635]: Soliciting pool server 2603:c020:8017:3eff::123 2026-03-06T22:14:57.829 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:14:57 ntpd[15662]: Soliciting pool server 185.125.190.57 2026-03-06T22:14:57.864 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:57 ntpd[15635]: Soliciting pool server 185.125.190.57 2026-03-06T22:14:59.887 INFO:teuthology.orchestra.run.vm08.stdout: 6 Mar 22:14:59 ntpd[15635]: ntpd: time slew +0.001791 s 2026-03-06T22:14:59.887 INFO:teuthology.orchestra.run.vm08.stdout:ntpd: time slew +0.001791s 2026-03-06T22:14:59.908 INFO:teuthology.orchestra.run.vm08.stdout: remote refid st t when poll reach delay offset jitter 2026-03-06T22:14:59.909 INFO:teuthology.orchestra.run.vm08.stdout:============================================================================== 2026-03-06T22:14:59.909 INFO:teuthology.orchestra.run.vm08.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:14:59.909 INFO:teuthology.orchestra.run.vm08.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:14:59.909 INFO:teuthology.orchestra.run.vm08.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:14:59.909 INFO:teuthology.orchestra.run.vm08.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:14:59.909 INFO:teuthology.orchestra.run.vm08.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:15:00.852 INFO:teuthology.orchestra.run.vm03.stdout: 6 Mar 22:15:00 ntpd[15662]: ntpd: time slew -0.000990 s 2026-03-06T22:15:00.853 INFO:teuthology.orchestra.run.vm03.stdout:ntpd: time slew -0.000990s 2026-03-06T22:15:00.872 INFO:teuthology.orchestra.run.vm03.stdout: remote refid st t when poll reach delay offset jitter 2026-03-06T22:15:00.872 INFO:teuthology.orchestra.run.vm03.stdout:============================================================================== 2026-03-06T22:15:00.872 INFO:teuthology.orchestra.run.vm03.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:15:00.872 INFO:teuthology.orchestra.run.vm03.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:15:00.872 INFO:teuthology.orchestra.run.vm03.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:15:00.872 INFO:teuthology.orchestra.run.vm03.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:15:00.872 INFO:teuthology.orchestra.run.vm03.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:15:00.872 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-06T22:15:00.921 INFO:tasks.cephadm:Config: {'roleless': True, 'conf': {'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'osd shutdown pgref assert': True}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'CEPHADM_DAEMON_PLACE_FAIL', 'CEPHADM_FAILED_DAEMON'], 'log-only-match': ['CEPHADM_'], 'sha1': '340d3c24fc6ae7529322dc7ccee6c6cb2589da0a', 'cephadm_binary_url': 'https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm', 'containers': {'image': 'harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5'}} 2026-03-06T22:15:00.921 INFO:tasks.cephadm:Provided image contains tag or digest, using it as is 2026-03-06T22:15:00.921 INFO:tasks.cephadm:Cluster image is harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 2026-03-06T22:15:00.921 INFO:tasks.cephadm:Cluster fsid is 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:15:00.921 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-06T22:15:00.921 INFO:tasks.cephadm:No mon roles; fabricating mons 2026-03-06T22:15:00.921 INFO:tasks.cephadm:Monitor IPs: {'mon.vm03': '192.168.123.103', 'mon.vm08': '192.168.123.108'} 2026-03-06T22:15:00.921 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-06T22:15:00.921 DEBUG:teuthology.orchestra.run.vm03:> sudo hostname $(hostname -s) 2026-03-06T22:15:00.928 DEBUG:teuthology.orchestra.run.vm08:> sudo hostname $(hostname -s) 2026-03-06T22:15:00.936 INFO:tasks.cephadm:Downloading cephadm from url: https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm 2026-03-06T22:15:00.936 DEBUG:teuthology.orchestra.run.vm03:> curl --silent -L https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-06T22:15:02.058 INFO:teuthology.orchestra.run.vm03.stdout:-rw-rw-r-- 1 ubuntu ubuntu 787672 Mar 6 22:15 /home/ubuntu/cephtest/cephadm 2026-03-06T22:15:02.058 DEBUG:teuthology.orchestra.run.vm08:> curl --silent -L https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-06T22:15:03.096 INFO:teuthology.orchestra.run.vm08.stdout:-rw-rw-r-- 1 ubuntu ubuntu 787672 Mar 6 22:15 /home/ubuntu/cephtest/cephadm 2026-03-06T22:15:03.096 DEBUG:teuthology.orchestra.run.vm03:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-06T22:15:03.100 DEBUG:teuthology.orchestra.run.vm08:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-06T22:15:03.108 INFO:tasks.cephadm:Pulling image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 on all hosts... 2026-03-06T22:15:03.108 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 pull 2026-03-06T22:15:03.143 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 pull 2026-03-06T22:15:03.392 INFO:teuthology.orchestra.run.vm08.stderr:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5... 2026-03-06T22:15:03.394 INFO:teuthology.orchestra.run.vm03.stderr:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5... 2026-03-06T22:15:25.229 INFO:teuthology.orchestra.run.vm08.stdout:{ 2026-03-06T22:15:25.229 INFO:teuthology.orchestra.run.vm08.stdout: "ceph_version": "ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable)", 2026-03-06T22:15:25.229 INFO:teuthology.orchestra.run.vm08.stdout: "image_id": "8bccc98d839aa18345ec1336292d0452ca331737e49f12524f635044dcabcfe1", 2026-03-06T22:15:25.229 INFO:teuthology.orchestra.run.vm08.stdout: "repo_digests": [ 2026-03-06T22:15:25.229 INFO:teuthology.orchestra.run.vm08.stdout: "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0" 2026-03-06T22:15:25.229 INFO:teuthology.orchestra.run.vm08.stdout: ] 2026-03-06T22:15:25.229 INFO:teuthology.orchestra.run.vm08.stdout:} 2026-03-06T22:15:25.476 INFO:teuthology.orchestra.run.vm03.stdout:{ 2026-03-06T22:15:25.476 INFO:teuthology.orchestra.run.vm03.stdout: "ceph_version": "ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable)", 2026-03-06T22:15:25.477 INFO:teuthology.orchestra.run.vm03.stdout: "image_id": "8bccc98d839aa18345ec1336292d0452ca331737e49f12524f635044dcabcfe1", 2026-03-06T22:15:25.477 INFO:teuthology.orchestra.run.vm03.stdout: "repo_digests": [ 2026-03-06T22:15:25.477 INFO:teuthology.orchestra.run.vm03.stdout: "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0" 2026-03-06T22:15:25.477 INFO:teuthology.orchestra.run.vm03.stdout: ] 2026-03-06T22:15:25.477 INFO:teuthology.orchestra.run.vm03.stdout:} 2026-03-06T22:15:25.504 DEBUG:teuthology.orchestra.run.vm03:> sudo mkdir -p /etc/ceph 2026-03-06T22:15:25.512 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /etc/ceph 2026-03-06T22:15:25.521 DEBUG:teuthology.orchestra.run.vm03:> sudo chmod 777 /etc/ceph 2026-03-06T22:15:25.562 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod 777 /etc/ceph 2026-03-06T22:15:25.573 INFO:tasks.cephadm:Writing seed config... 2026-03-06T22:15:25.573 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-06T22:15:25.573 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-06T22:15:25.573 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-06T22:15:25.573 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-06T22:15:25.573 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-06T22:15:25.573 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-06T22:15:25.573 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-06T22:15:25.573 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-06T22:15:25.573 INFO:tasks.cephadm: override: [osd] osd shutdown pgref assert = True 2026-03-06T22:15:25.574 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-06T22:15:25.574 DEBUG:teuthology.orchestra.run.vm03:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-06T22:15:25.608 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 894e000c-19a1-11f1-8dbe-23b24380a082 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = True bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-06T22:15:25.608 DEBUG:teuthology.orchestra.run.vm03:mon.vm03> sudo journalctl -f -n 0 -u ceph-894e000c-19a1-11f1-8dbe-23b24380a082@mon.vm03.service 2026-03-06T22:15:25.650 INFO:tasks.cephadm:Bootstrapping... 2026-03-06T22:15:25.650 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 -v bootstrap --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-ip 192.168.123.103 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-06T22:15:25.954 INFO:teuthology.orchestra.run.vm03.stdout:-------------------------------------------------------------------------------- 2026-03-06T22:15:25.954 INFO:teuthology.orchestra.run.vm03.stdout:cephadm ['--image', 'harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5', '-v', 'bootstrap', '--fsid', '894e000c-19a1-11f1-8dbe-23b24380a082', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-ip', '192.168.123.103', '--skip-admin-label'] 2026-03-06T22:15:25.954 INFO:teuthology.orchestra.run.vm03.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-06T22:15:25.954 INFO:teuthology.orchestra.run.vm03.stdout:Verifying podman|docker is present... 2026-03-06T22:15:25.954 INFO:teuthology.orchestra.run.vm03.stdout:Verifying lvm2 is present... 2026-03-06T22:15:25.954 INFO:teuthology.orchestra.run.vm03.stdout:Verifying time synchronization is in place... 2026-03-06T22:15:25.957 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-06T22:15:25.957 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-06T22:15:25.960 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-06T22:15:25.960 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout inactive 2026-03-06T22:15:25.962 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl is-enabled chronyd.service 2026-03-06T22:15:25.962 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Failed to get unit file state for chronyd.service: No such file or directory 2026-03-06T22:15:25.964 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 3 from systemctl is-active chronyd.service 2026-03-06T22:15:25.964 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout inactive 2026-03-06T22:15:25.967 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl is-enabled systemd-timesyncd.service 2026-03-06T22:15:25.967 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout masked 2026-03-06T22:15:25.969 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 3 from systemctl is-active systemd-timesyncd.service 2026-03-06T22:15:25.969 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout inactive 2026-03-06T22:15:25.971 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl is-enabled ntpd.service 2026-03-06T22:15:25.971 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Failed to get unit file state for ntpd.service: No such file or directory 2026-03-06T22:15:25.973 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 3 from systemctl is-active ntpd.service 2026-03-06T22:15:25.973 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout inactive 2026-03-06T22:15:25.977 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout enabled 2026-03-06T22:15:25.979 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout active 2026-03-06T22:15:25.979 INFO:teuthology.orchestra.run.vm03.stdout:Unit ntp.service is enabled and running 2026-03-06T22:15:25.979 INFO:teuthology.orchestra.run.vm03.stdout:Repeating the final host check... 2026-03-06T22:15:25.979 INFO:teuthology.orchestra.run.vm03.stdout:docker (/usr/bin/docker) is present 2026-03-06T22:15:25.979 INFO:teuthology.orchestra.run.vm03.stdout:systemctl is present 2026-03-06T22:15:25.979 INFO:teuthology.orchestra.run.vm03.stdout:lvcreate is present 2026-03-06T22:15:25.982 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-06T22:15:25.982 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-06T22:15:25.986 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-06T22:15:25.986 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout inactive 2026-03-06T22:15:25.988 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl is-enabled chronyd.service 2026-03-06T22:15:25.988 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Failed to get unit file state for chronyd.service: No such file or directory 2026-03-06T22:15:25.991 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 3 from systemctl is-active chronyd.service 2026-03-06T22:15:25.991 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout inactive 2026-03-06T22:15:25.993 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl is-enabled systemd-timesyncd.service 2026-03-06T22:15:25.993 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout masked 2026-03-06T22:15:25.996 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 3 from systemctl is-active systemd-timesyncd.service 2026-03-06T22:15:25.996 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout inactive 2026-03-06T22:15:25.999 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl is-enabled ntpd.service 2026-03-06T22:15:25.999 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Failed to get unit file state for ntpd.service: No such file or directory 2026-03-06T22:15:26.002 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 3 from systemctl is-active ntpd.service 2026-03-06T22:15:26.002 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout inactive 2026-03-06T22:15:26.005 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout enabled 2026-03-06T22:15:26.008 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout active 2026-03-06T22:15:26.008 INFO:teuthology.orchestra.run.vm03.stdout:Unit ntp.service is enabled and running 2026-03-06T22:15:26.008 INFO:teuthology.orchestra.run.vm03.stdout:Host looks OK 2026-03-06T22:15:26.008 INFO:teuthology.orchestra.run.vm03.stdout:Cluster fsid: 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:15:26.008 INFO:teuthology.orchestra.run.vm03.stdout:Acquiring lock 140449292339376 on /run/cephadm/894e000c-19a1-11f1-8dbe-23b24380a082.lock 2026-03-06T22:15:26.008 INFO:teuthology.orchestra.run.vm03.stdout:Lock 140449292339376 acquired on /run/cephadm/894e000c-19a1-11f1-8dbe-23b24380a082.lock 2026-03-06T22:15:26.008 INFO:teuthology.orchestra.run.vm03.stdout:Verifying IP 192.168.123.103 port 3300 ... 2026-03-06T22:15:26.008 INFO:teuthology.orchestra.run.vm03.stdout:Verifying IP 192.168.123.103 port 6789 ... 2026-03-06T22:15:26.008 INFO:teuthology.orchestra.run.vm03.stdout:Base mon IP(s) is [192.168.123.103:3300, 192.168.123.103:6789], mon addrv is [v2:192.168.123.103:3300,v1:192.168.123.103:6789] 2026-03-06T22:15:26.011 INFO:teuthology.orchestra.run.vm03.stdout:/usr/sbin/ip: stdout default via 192.168.123.1 dev ens3 proto dhcp src 192.168.123.103 metric 100 2026-03-06T22:15:26.011 INFO:teuthology.orchestra.run.vm03.stdout:/usr/sbin/ip: stdout 172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.0.1 linkdown 2026-03-06T22:15:26.011 INFO:teuthology.orchestra.run.vm03.stdout:/usr/sbin/ip: stdout 192.168.123.0/24 dev ens3 proto kernel scope link src 192.168.123.103 metric 100 2026-03-06T22:15:26.011 INFO:teuthology.orchestra.run.vm03.stdout:/usr/sbin/ip: stdout 192.168.123.1 dev ens3 proto dhcp scope link src 192.168.123.103 metric 100 2026-03-06T22:15:26.012 INFO:teuthology.orchestra.run.vm03.stdout:/usr/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-06T22:15:26.012 INFO:teuthology.orchestra.run.vm03.stdout:/usr/sbin/ip: stdout fe80::/64 dev ens3 proto kernel metric 256 pref medium 2026-03-06T22:15:26.014 INFO:teuthology.orchestra.run.vm03.stdout:/usr/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-06T22:15:26.014 INFO:teuthology.orchestra.run.vm03.stdout:/usr/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-06T22:15:26.014 INFO:teuthology.orchestra.run.vm03.stdout:/usr/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-06T22:15:26.014 INFO:teuthology.orchestra.run.vm03.stdout:/usr/sbin/ip: stdout 2: ens3: mtu 1500 state UP qlen 1000 2026-03-06T22:15:26.014 INFO:teuthology.orchestra.run.vm03.stdout:/usr/sbin/ip: stdout inet6 fe80::5055:ff:fe00:3/64 scope link 2026-03-06T22:15:26.014 INFO:teuthology.orchestra.run.vm03.stdout:/usr/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-06T22:15:26.014 INFO:teuthology.orchestra.run.vm03.stdout:Mon IP `192.168.123.103` is in CIDR network `192.168.123.0/24` 2026-03-06T22:15:26.015 INFO:teuthology.orchestra.run.vm03.stdout:Mon IP `192.168.123.103` is in CIDR network `192.168.123.0/24` 2026-03-06T22:15:26.015 INFO:teuthology.orchestra.run.vm03.stdout:Mon IP `192.168.123.103` is in CIDR network `192.168.123.1/32` 2026-03-06T22:15:26.015 INFO:teuthology.orchestra.run.vm03.stdout:Mon IP `192.168.123.103` is in CIDR network `192.168.123.1/32` 2026-03-06T22:15:26.015 INFO:teuthology.orchestra.run.vm03.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24', '192.168.123.1/32', '192.168.123.1/32'] 2026-03-06T22:15:26.015 INFO:teuthology.orchestra.run.vm03.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-06T22:15:26.015 INFO:teuthology.orchestra.run.vm03.stdout:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5... 2026-03-06T22:15:26.726 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/docker: stdout cobaltcore-storage-v19.2.3-fasttrack-5: Pulling from custom-ceph/ceph/ceph 2026-03-06T22:15:26.726 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/docker: stdout Digest: sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0 2026-03-06T22:15:26.726 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/docker: stdout Status: Image is up to date for harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 2026-03-06T22:15:26.726 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/docker: stdout harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 2026-03-06T22:15:27.089 INFO:teuthology.orchestra.run.vm03.stdout:ceph: stdout ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable) 2026-03-06T22:15:27.089 INFO:teuthology.orchestra.run.vm03.stdout:Ceph version: ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable) 2026-03-06T22:15:27.089 INFO:teuthology.orchestra.run.vm03.stdout:Extracting ceph user uid/gid from container image... 2026-03-06T22:15:27.231 INFO:teuthology.orchestra.run.vm03.stdout:stat: stdout 167 167 2026-03-06T22:15:27.231 INFO:teuthology.orchestra.run.vm03.stdout:Creating initial keys... 2026-03-06T22:15:27.425 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-authtool: stdout AQDvQ6tplVyyFBAANAur51nOLlUr8d9aS8rlBw== 2026-03-06T22:15:27.572 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-authtool: stdout AQDvQ6tpVMvJHxAAAcK+9wSC1eUUCvCO/2YODg== 2026-03-06T22:15:27.680 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-authtool: stdout AQDvQ6tpIQK6JhAAy5V3u2Hp4C3FK1nxXrBgzg== 2026-03-06T22:15:27.680 INFO:teuthology.orchestra.run.vm03.stdout:Creating initial monmap... 2026-03-06T22:15:27.792 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-06T22:15:27.792 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-06T22:15:27.792 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:15:27.792 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-06T22:15:27.792 INFO:teuthology.orchestra.run.vm03.stdout:monmaptool for vm03 [v2:192.168.123.103:3300,v1:192.168.123.103:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-06T22:15:27.792 INFO:teuthology.orchestra.run.vm03.stdout:setting min_mon_release = quincy 2026-03-06T22:15:27.792 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: set fsid to 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:15:27.792 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-06T22:15:27.792 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:15:27.792 INFO:teuthology.orchestra.run.vm03.stdout:Creating mon... 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.884+0000 7f48f073bd80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.884+0000 7f48f073bd80 1 imported monmap: 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr epoch 0 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr fsid 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr last_changed 2026-03-06T21:15:27.758447+0000 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr created 2026-03-06T21:15:27.758447+0000 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr min_mon_release 17 (quincy) 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr election_strategy: 1 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.vm03 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.884+0000 7f48f073bd80 0 /usr/bin/ceph-mon: set fsid to 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: RocksDB version: 7.9.2 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Git sha 0 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Compile date 2026-03-06 13:52:12 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: DB SUMMARY 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: DB Session ID: ZYNX0DS6ST7KK3LWKLQX 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-vm03/store.db dir, Total Num: 0, files: 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-vm03/store.db: 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.error_if_exists: 0 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.create_if_missing: 1 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.paranoid_checks: 1 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.env: 0x5595fab0dca0 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.info_log: 0x559605478ce0 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.statistics: (nil) 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.use_fsync: 0 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.max_log_file_size: 0 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.allow_fallocate: 1 2026-03-06T22:15:28.066 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.use_direct_reads: 0 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.db_log_dir: 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.wal_dir: 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.write_buffer_manager: 0x55960546f5e0 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.unordered_write: 0 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.row_cache: None 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.wal_filter: None 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.two_write_queues: 0 2026-03-06T22:15:28.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.wal_compression: 0 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.atomic_flush: 0 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.log_readahead_size: 0 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.max_background_jobs: 2 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.max_background_compactions: -1 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.max_subcompactions: 1 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-06T22:15:28.069 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.max_open_files: -1 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Options.max_background_flushes: -1 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Compression algorithms supported: 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: kZSTD supported: 0 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: kXpressCompression supported: 0 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: kBZip2Compression supported: 0 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: kLZ4Compression supported: 1 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: kZlibCompression supported: 1 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: kSnappyCompression supported: 1 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.888+0000 7f48f073bd80 4 rocksdb: [db/db_impl/db_impl_open.cc:317] Creating manifest 1 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-vm03/store.db/MANIFEST-000001 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.merge_operator: 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compaction_filter: None 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55960546b400) 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr cache_index_and_filter_blocks: 1 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr cache_index_and_filter_blocks_with_high_priority: 0 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr pin_top_level_index_and_filter: 1 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr index_type: 0 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr data_block_index_type: 0 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr index_shortening: 1 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr data_block_hash_table_util_ratio: 0.750000 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr checksum: 4 2026-03-06T22:15:28.070 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr no_block_cache: 0 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr block_cache: 0x5596054911f0 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr block_cache_name: BinnedLRUCache 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr block_cache_options: 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr capacity : 536870912 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr num_shard_bits : 4 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr strict_capacity_limit : 0 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr high_pri_pool_ratio: 0.000 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr block_cache_compressed: (nil) 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr persistent_cache: (nil) 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr block_size: 4096 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr block_size_deviation: 10 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr block_restart_interval: 16 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr index_block_restart_interval: 1 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr metadata_block_size: 4096 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr partition_filters: 0 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr use_delta_encoding: 1 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr filter_policy: bloomfilter 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr whole_key_filtering: 1 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr verify_compression: 0 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr read_amp_bytes_per_bit: 0 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr format_version: 5 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr enable_index_compression: 1 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr block_align: 0 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr max_auto_readahead_size: 262144 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr prepopulate_block_cache: 0 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr initial_auto_readahead_size: 8192 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr num_file_reads_for_auto_readahead: 2 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compression: NoCompression 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.num_levels: 7 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-06T22:15:28.071 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-06T22:15:28.072 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.inplace_update_support: 0 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.bloom_locality: 0 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.max_successive_merges: 0 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.ttl: 2592000 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.enable_blob_files: false 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.min_blob_size: 0 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-vm03/store.db/MANIFEST-000001 succeeded,manifest_file_number is 1, next_file_number is 3, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 1371237a-4543-4b07-9dc4-cfdb64579b14 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.892+0000 7f48f073bd80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 5 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.896+0000 7f48f073bd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x559605492e00 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.896+0000 7f48f073bd80 4 rocksdb: DB pointer 0x559605576000 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.896+0000 7f48e7ec5640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.896+0000 7f48e7ec5640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr ** DB Stats ** 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T22:15:28.073 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr ** Compaction Stats [default] ** 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr ** Compaction Stats [default] ** 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Flush(GB): cumulative 0.000, interval 0.000 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr AddFile(GB): cumulative 0.000, interval 0.000 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr AddFile(Total Files): cumulative 0, interval 0 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr AddFile(L0 Files): cumulative 0, interval 0 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr AddFile(Keys): cumulative 0, interval 0 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Block cache BinnedLRUCache@0x5596054911f0#7 capacity: 512.00 MB usage: 0.00 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1e-05 secs_since: 0 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr Block cache entry stats(count,size,portion): Misc(1,0.00 KB,0%) 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr ** File Read Latency Histogram By Level [default] ** 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.904+0000 7f48f073bd80 4 rocksdb: [db/db_impl/db_impl.cc:496] Shutdown: canceling all background work 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.904+0000 7f48f073bd80 4 rocksdb: [db/db_impl/db_impl.cc:704] Shutdown complete 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-06T21:15:27.904+0000 7f48f073bd80 0 /usr/bin/ceph-mon: created monfs at /var/lib/ceph/mon/ceph-vm03 for mon.vm03 2026-03-06T22:15:28.074 INFO:teuthology.orchestra.run.vm03.stdout:create mon.vm03 on 2026-03-06T22:15:28.516 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-06T22:15:28.773 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-894e000c-19a1-11f1-8dbe-23b24380a082.target → /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082.target. 2026-03-06T22:15:28.773 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-894e000c-19a1-11f1-8dbe-23b24380a082.target → /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082.target. 2026-03-06T22:15:28.988 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-894e000c-19a1-11f1-8dbe-23b24380a082@mon.vm03 2026-03-06T22:15:28.988 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Failed to reset failed state of unit ceph-894e000c-19a1-11f1-8dbe-23b24380a082@mon.vm03.service: Unit ceph-894e000c-19a1-11f1-8dbe-23b24380a082@mon.vm03.service not loaded. 2026-03-06T22:15:29.181 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082.target.wants/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@mon.vm03.service → /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service. 2026-03-06T22:15:29.189 INFO:teuthology.orchestra.run.vm03.stdout:firewalld does not appear to be present 2026-03-06T22:15:29.189 INFO:teuthology.orchestra.run.vm03.stdout:Not possible to enable service . firewalld.service is not available 2026-03-06T22:15:29.189 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mon to start... 2026-03-06T22:15:29.189 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mon... 2026-03-06T22:15:29.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:29 vm03 bash[16576]: cluster 2026-03-06T21:15:29.307766+0000 mon.vm03 (mon.0) 1 : cluster [INF] mon.vm03 is new leader, mons vm03 in quorum (ranks 0) 2026-03-06T22:15:29.549 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout cluster: 2026-03-06T22:15:29.549 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout id: 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:15:29.549 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-06T22:15:29.549 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-06T22:15:29.549 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout services: 2026-03-06T22:15:29.549 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum vm03 (age 0.178024s) 2026-03-06T22:15:29.549 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-06T22:15:29.549 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-06T22:15:29.549 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-06T22:15:29.549 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout data: 2026-03-06T22:15:29.549 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-06T22:15:29.549 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-06T22:15:29.549 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-06T22:15:29.549 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout pgs: 2026-03-06T22:15:29.549 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-06T22:15:29.549 INFO:teuthology.orchestra.run.vm03.stdout:mon is available 2026-03-06T22:15:29.549 INFO:teuthology.orchestra.run.vm03.stdout:Assimilating anything we can from ceph.conf... 2026-03-06T22:15:30.178 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-06T22:15:30.178 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [global] 2026-03-06T22:15:30.178 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout fsid = 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:15:30.178 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-06T22:15:30.178 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.103:3300,v1:192.168.123.103:6789] 2026-03-06T22:15:30.178 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-06T22:15:30.178 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-06T22:15:30.178 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-06T22:15:30.178 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-06T22:15:30.178 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-06T22:15:30.178 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-06T22:15:30.178 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-06T22:15:30.179 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-06T22:15:30.179 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [osd] 2026-03-06T22:15:30.179 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-06T22:15:30.179 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-06T22:15:30.179 INFO:teuthology.orchestra.run.vm03.stdout:Generating new minimal ceph.conf... 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.312304+0000 mon.vm03 (mon.0) 2 : cluster [INF] mon.vm03 is new leader, mons vm03 in quorum (ranks 0) 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.312304+0000 mon.vm03 (mon.0) 2 : cluster [INF] mon.vm03 is new leader, mons vm03 in quorum (ranks 0) 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.312796+0000 mon.vm03 (mon.0) 3 : cluster [DBG] monmap epoch 1 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.312796+0000 mon.vm03 (mon.0) 3 : cluster [DBG] monmap epoch 1 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.312803+0000 mon.vm03 (mon.0) 4 : cluster [DBG] fsid 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.312803+0000 mon.vm03 (mon.0) 4 : cluster [DBG] fsid 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.312806+0000 mon.vm03 (mon.0) 5 : cluster [DBG] last_changed 2026-03-06T21:15:27.758447+0000 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.312806+0000 mon.vm03 (mon.0) 5 : cluster [DBG] last_changed 2026-03-06T21:15:27.758447+0000 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.312810+0000 mon.vm03 (mon.0) 6 : cluster [DBG] created 2026-03-06T21:15:27.758447+0000 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.312810+0000 mon.vm03 (mon.0) 6 : cluster [DBG] created 2026-03-06T21:15:27.758447+0000 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.312813+0000 mon.vm03 (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.312813+0000 mon.vm03 (mon.0) 7 : cluster [DBG] min_mon_release 19 (squid) 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.312817+0000 mon.vm03 (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.312817+0000 mon.vm03 (mon.0) 8 : cluster [DBG] election_strategy: 1 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.312820+0000 mon.vm03 (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.vm03 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.312820+0000 mon.vm03 (mon.0) 9 : cluster [DBG] 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.vm03 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.315235+0000 mon.vm03 (mon.0) 10 : cluster [DBG] fsmap 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.315235+0000 mon.vm03 (mon.0) 10 : cluster [DBG] fsmap 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.320557+0000 mon.vm03 (mon.0) 11 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.320557+0000 mon.vm03 (mon.0) 11 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.322317+0000 mon.vm03 (mon.0) 12 : cluster [DBG] mgrmap e1: no daemons active 2026-03-06T22:15:30.611 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: cluster 2026-03-06T21:15:29.322317+0000 mon.vm03 (mon.0) 12 : cluster [DBG] mgrmap e1: no daemons active 2026-03-06T22:15:30.612 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: audit 2026-03-06T21:15:29.490262+0000 mon.vm03 (mon.0) 13 : audit [DBG] from='client.? 192.168.123.103:0/3475049104' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-06T22:15:30.612 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: audit 2026-03-06T21:15:29.490262+0000 mon.vm03 (mon.0) 13 : audit [DBG] from='client.? 192.168.123.103:0/3475049104' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-06T22:15:30.612 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: audit 2026-03-06T21:15:29.809756+0000 mon.vm03 (mon.0) 14 : audit [INF] from='client.? 192.168.123.103:0/3821341844' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-06T22:15:30.612 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: audit 2026-03-06T21:15:29.809756+0000 mon.vm03 (mon.0) 14 : audit [INF] from='client.? 192.168.123.103:0/3821341844' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-06T22:15:30.612 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: audit 2026-03-06T21:15:29.816698+0000 mon.vm03 (mon.0) 15 : audit [INF] from='client.? 192.168.123.103:0/3821341844' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-06T22:15:30.612 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: audit 2026-03-06T21:15:29.816698+0000 mon.vm03 (mon.0) 15 : audit [INF] from='client.? 192.168.123.103:0/3821341844' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-06T22:15:30.646 INFO:teuthology.orchestra.run.vm03.stdout:Restarting the monitor... 2026-03-06T22:15:30.762 INFO:teuthology.orchestra.run.vm03.stdout:Setting public_network to 192.168.123.1/32,192.168.123.0/24 in mon config section 2026-03-06T22:15:30.865 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 systemd[1]: Stopping Ceph mon.vm03 for 894e000c-19a1-11f1-8dbe-23b24380a082... 2026-03-06T22:15:30.865 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: debug 2026-03-06T21:15:30.680+0000 7fd77b447640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.vm03 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-06T22:15:30.865 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16576]: debug 2026-03-06T21:15:30.680+0000 7fd77b447640 -1 mon.vm03@0(leader) e1 *** Got Signal Terminated *** 2026-03-06T22:15:30.865 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[16968]: ceph-894e000c-19a1-11f1-8dbe-23b24380a082-mon-vm03 2026-03-06T22:15:30.865 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 systemd[1]: ceph-894e000c-19a1-11f1-8dbe-23b24380a082@mon.vm03.service: Deactivated successfully. 2026-03-06T22:15:30.865 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 systemd[1]: Stopped Ceph mon.vm03 for 894e000c-19a1-11f1-8dbe-23b24380a082. 2026-03-06T22:15:30.865 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 systemd[1]: Started Ceph mon.vm03 for 894e000c-19a1-11f1-8dbe-23b24380a082. 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.892+0000 7ff81d271d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.892+0000 7ff81d271d80 0 ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable), process ceph-mon, pid 7 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.892+0000 7ff81d271d80 0 pidfile_write: ignore empty --pid-file 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.892+0000 7ff81d271d80 0 load: jerasure load: lrc 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Git sha 0 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Compile date 2026-03-06 13:52:12 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: DB SUMMARY 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: DB Session ID: 6SVAB5TPXV61HHK1YKQY 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: CURRENT file: CURRENT 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-vm03/store.db dir, Total Num: 1, files: 000008.sst 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-vm03/store.db: 000009.log size: 86821 ; 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.error_if_exists: 0 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.create_if_missing: 0 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-06T22:15:31.148 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.env: 0x564c3f139ca0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.info_log: 0x564c7b488500 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.statistics: (nil) 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.use_fsync: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.db_log_dir: 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.wal_dir: 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.write_buffer_manager: 0x564c7b48d900 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.unordered_write: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.row_cache: None 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.wal_filter: None 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-06T22:15:31.149 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.two_write_queues: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.wal_compression: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.atomic_flush: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_open_files: -1 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Compression algorithms supported: 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: kZSTD supported: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: kXpressCompression supported: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: kZlibCompression supported: 1 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-vm03/store.db/MANIFEST-000010 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.merge_operator: 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compaction_filter: None 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x564c7b4884c0) 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cache_index_and_filter_blocks: 1 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: pin_top_level_index_and_filter: 1 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: index_type: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: data_block_index_type: 0 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: index_shortening: 1 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: data_block_hash_table_util_ratio: 0.750000 2026-03-06T22:15:31.150 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: checksum: 4 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: no_block_cache: 0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: block_cache: 0x564c7b4af1f0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: block_cache_name: BinnedLRUCache 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: block_cache_options: 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: capacity : 536870912 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: num_shard_bits : 4 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: strict_capacity_limit : 0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: high_pri_pool_ratio: 0.000 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: block_cache_compressed: (nil) 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: persistent_cache: (nil) 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: block_size: 4096 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: block_size_deviation: 10 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: block_restart_interval: 16 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: index_block_restart_interval: 1 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: metadata_block_size: 4096 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: partition_filters: 0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: use_delta_encoding: 1 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: filter_policy: bloomfilter 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: whole_key_filtering: 1 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: verify_compression: 0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: read_amp_bytes_per_bit: 0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: format_version: 5 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: enable_index_compression: 1 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: block_align: 0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: max_auto_readahead_size: 262144 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: prepopulate_block_cache: 0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: initial_auto_readahead_size: 8192 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: num_file_reads_for_auto_readahead: 2 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compression: NoCompression 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.num_levels: 7 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-06T22:15:31.151 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.bloom_locality: 0 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.ttl: 2592000 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.enable_blob_files: false 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.min_blob_size: 0 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-06T22:15:31.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.896+0000 7ff81d271d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.904+0000 7ff81d271d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-vm03/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.904+0000 7ff81d271d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.904+0000 7ff81d271d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 1371237a-4543-4b07-9dc4-cfdb64579b14 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.904+0000 7ff81d271d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772831730912126, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.904+0000 7ff81d271d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.908+0000 7ff81d271d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772831730914873, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 83798, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 243, "table_properties": {"data_size": 81967, "index_size": 220, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 581, "raw_key_size": 9991, "raw_average_key_size": 47, "raw_value_size": 76276, "raw_average_value_size": 363, "num_data_blocks": 11, "num_entries": 210, "num_filter_entries": 210, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1772831730, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "1371237a-4543-4b07-9dc4-cfdb64579b14", "db_session_id": "6SVAB5TPXV61HHK1YKQY", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.908+0000 7ff81d271d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772831730914980, "job": 1, "event": "recovery_finished"} 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.908+0000 7ff81d271d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.912+0000 7ff81d271d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-vm03/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.912+0000 7ff81d271d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x564c7b4b0e00 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.912+0000 7ff81d271d80 4 rocksdb: DB pointer 0x564c7b5cc000 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.912+0000 7ff81303b640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: debug 2026-03-06T21:15:30.912+0000 7ff81303b640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: ** DB Stats ** 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: ** Compaction Stats [default] ** 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: L0 2/0 83.71 KB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 32.9 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Sum 2/0 83.71 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 32.9 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 32.9 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: ** Compaction Stats [default] ** 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 32.9 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: AddFile(Total Files): cumulative 0, interval 0 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: AddFile(Keys): cumulative 0, interval 0 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Cumulative compaction: 0.00 GB write, 4.66 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Interval compaction: 0.00 GB write, 4.66 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Block cache BinnedLRUCache@0x564c7b4af1f0#7 capacity: 512.00 MB usage: 1.17 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1.1e-05 secs_since: 0 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: Block cache entry stats(count,size,portion): FilterBlock(2,0.77 KB,0.000146031%) IndexBlock(2,0.41 KB,7.7486e-05%) Misc(1,0.00 KB,0%) 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: ** File Read Latency Histogram By Level [default] ** 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926339+0000 mon.vm03 (mon.0) 1 : cluster [INF] mon.vm03 is new leader, mons vm03 in quorum (ranks 0) 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926339+0000 mon.vm03 (mon.0) 1 : cluster [INF] mon.vm03 is new leader, mons vm03 in quorum (ranks 0) 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926390+0000 mon.vm03 (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926390+0000 mon.vm03 (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926396+0000 mon.vm03 (mon.0) 3 : cluster [DBG] fsid 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926396+0000 mon.vm03 (mon.0) 3 : cluster [DBG] fsid 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:15:31.153 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926400+0000 mon.vm03 (mon.0) 4 : cluster [DBG] last_changed 2026-03-06T21:15:27.758447+0000 2026-03-06T22:15:31.154 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926400+0000 mon.vm03 (mon.0) 4 : cluster [DBG] last_changed 2026-03-06T21:15:27.758447+0000 2026-03-06T22:15:31.154 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926408+0000 mon.vm03 (mon.0) 5 : cluster [DBG] created 2026-03-06T21:15:27.758447+0000 2026-03-06T22:15:31.154 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926408+0000 mon.vm03 (mon.0) 5 : cluster [DBG] created 2026-03-06T21:15:27.758447+0000 2026-03-06T22:15:31.154 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926412+0000 mon.vm03 (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-06T22:15:31.154 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926412+0000 mon.vm03 (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-06T22:15:31.154 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926416+0000 mon.vm03 (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-06T22:15:31.154 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926416+0000 mon.vm03 (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-06T22:15:31.154 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926421+0000 mon.vm03 (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.vm03 2026-03-06T22:15:31.154 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926421+0000 mon.vm03 (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.vm03 2026-03-06T22:15:31.154 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926684+0000 mon.vm03 (mon.0) 9 : cluster [DBG] fsmap 2026-03-06T22:15:31.154 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926684+0000 mon.vm03 (mon.0) 9 : cluster [DBG] fsmap 2026-03-06T22:15:31.154 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926700+0000 mon.vm03 (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-06T22:15:31.154 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.926700+0000 mon.vm03 (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-06T22:15:31.154 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.927374+0000 mon.vm03 (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-06T22:15:31.154 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:30 vm03 bash[17055]: cluster 2026-03-06T21:15:30.927374+0000 mon.vm03 (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-06T22:15:31.191 INFO:teuthology.orchestra.run.vm03.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-06T22:15:31.192 INFO:teuthology.orchestra.run.vm03.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-06T22:15:31.192 INFO:teuthology.orchestra.run.vm03.stdout:Creating mgr... 2026-03-06T22:15:31.192 INFO:teuthology.orchestra.run.vm03.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-06T22:15:31.192 INFO:teuthology.orchestra.run.vm03.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-06T22:15:31.192 INFO:teuthology.orchestra.run.vm03.stdout:Verifying port 0.0.0.0:8443 ... 2026-03-06T22:15:31.366 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-894e000c-19a1-11f1-8dbe-23b24380a082@mgr.vm03.uwuzgl 2026-03-06T22:15:31.366 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Failed to reset failed state of unit ceph-894e000c-19a1-11f1-8dbe-23b24380a082@mgr.vm03.uwuzgl.service: Unit ceph-894e000c-19a1-11f1-8dbe-23b24380a082@mgr.vm03.uwuzgl.service not loaded. 2026-03-06T22:15:31.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:31 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:15:31.525 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082.target.wants/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@mgr.vm03.uwuzgl.service → /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service. 2026-03-06T22:15:31.536 INFO:teuthology.orchestra.run.vm03.stdout:firewalld does not appear to be present 2026-03-06T22:15:31.536 INFO:teuthology.orchestra.run.vm03.stdout:Not possible to enable service . firewalld.service is not available 2026-03-06T22:15:31.536 INFO:teuthology.orchestra.run.vm03.stdout:firewalld does not appear to be present 2026-03-06T22:15:31.536 INFO:teuthology.orchestra.run.vm03.stdout:Not possible to open ports <[9283, 8765, 8443]>. firewalld.service is not available 2026-03-06T22:15:31.536 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mgr to start... 2026-03-06T22:15:31.536 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mgr... 2026-03-06T22:15:31.853 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:31 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsid": "894e000c-19a1-11f1-8dbe-23b24380a082", 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 0 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "vm03" 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_age": 0, 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T22:15:31.899 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T21:15:29:313048+0000", 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T21:15:29.313859+0000", 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-06T22:15:31.900 INFO:teuthology.orchestra.run.vm03.stdout:mgr not available, waiting (1/15)... 2026-03-06T22:15:32.419 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:32 vm03 bash[17055]: audit 2026-03-06T21:15:31.127496+0000 mon.vm03 (mon.0) 12 : audit [INF] from='client.? 192.168.123.103:0/1617127230' entity='client.admin' 2026-03-06T22:15:32.419 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:32 vm03 bash[17055]: audit 2026-03-06T21:15:31.127496+0000 mon.vm03 (mon.0) 12 : audit [INF] from='client.? 192.168.123.103:0/1617127230' entity='client.admin' 2026-03-06T22:15:32.419 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:32 vm03 bash[17055]: audit 2026-03-06T21:15:31.834278+0000 mon.vm03 (mon.0) 13 : audit [DBG] from='client.? 192.168.123.103:0/146187420' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T22:15:32.419 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:32 vm03 bash[17055]: audit 2026-03-06T21:15:31.834278+0000 mon.vm03 (mon.0) 13 : audit [DBG] from='client.? 192.168.123.103:0/146187420' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T22:15:34.276 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsid": "894e000c-19a1-11f1-8dbe-23b24380a082", 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 0 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "vm03" 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_age": 3, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T21:15:29:313048+0000", 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T22:15:34.277 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T22:15:34.278 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T22:15:34.278 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T22:15:34.278 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T22:15:34.278 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T22:15:34.278 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:34.278 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T22:15:34.278 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:34.278 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T22:15:34.278 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:34.278 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T21:15:29.313859+0000", 2026-03-06T22:15:34.278 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T22:15:34.278 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:34.278 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T22:15:34.278 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-06T22:15:34.278 INFO:teuthology.orchestra.run.vm03.stdout:mgr not available, waiting (2/15)... 2026-03-06T22:15:34.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:34 vm03 bash[17055]: audit 2026-03-06T21:15:34.200751+0000 mon.vm03 (mon.0) 14 : audit [DBG] from='client.? 192.168.123.103:0/2456822607' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T22:15:34.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:34 vm03 bash[17055]: audit 2026-03-06T21:15:34.200751+0000 mon.vm03 (mon.0) 14 : audit [DBG] from='client.? 192.168.123.103:0/2456822607' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T22:15:36.638 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-06T22:15:36.638 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-06T22:15:36.638 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsid": "894e000c-19a1-11f1-8dbe-23b24380a082", 2026-03-06T22:15:36.638 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T22:15:36.638 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T22:15:36.638 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T22:15:36.638 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T22:15:36.638 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:36.638 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T22:15:36.638 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T22:15:36.638 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 0 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "vm03" 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T21:15:29:313048+0000", 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T21:15:29.313859+0000", 2026-03-06T22:15:36.639 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T22:15:36.640 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:36.640 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T22:15:36.640 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-06T22:15:36.640 INFO:teuthology.orchestra.run.vm03.stdout:mgr not available, waiting (3/15)... 2026-03-06T22:15:36.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:36 vm03 bash[17055]: audit 2026-03-06T21:15:36.570679+0000 mon.vm03 (mon.0) 15 : audit [DBG] from='client.? 192.168.123.103:0/885182652' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T22:15:36.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:36 vm03 bash[17055]: audit 2026-03-06T21:15:36.570679+0000 mon.vm03 (mon.0) 15 : audit [DBG] from='client.? 192.168.123.103:0/885182652' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T22:15:39.043 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsid": "894e000c-19a1-11f1-8dbe-23b24380a082", 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 0 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "vm03" 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_age": 8, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:39.044 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T21:15:29:313048+0000", 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T21:15:29.313859+0000", 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-06T22:15:39.045 INFO:teuthology.orchestra.run.vm03.stdout:mgr not available, waiting (4/15)... 2026-03-06T22:15:39.374 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:39 vm03 bash[17055]: audit 2026-03-06T21:15:38.954496+0000 mon.vm03 (mon.0) 16 : audit [DBG] from='client.? 192.168.123.103:0/3939997348' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T22:15:39.374 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:39 vm03 bash[17055]: audit 2026-03-06T21:15:38.954496+0000 mon.vm03 (mon.0) 16 : audit [DBG] from='client.? 192.168.123.103:0/3939997348' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsid": "894e000c-19a1-11f1-8dbe-23b24380a082", 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 0 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "vm03" 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_age": 10, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T22:15:41.436 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T21:15:29:313048+0000", 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T21:15:29.313859+0000", 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-06T22:15:41.437 INFO:teuthology.orchestra.run.vm03.stdout:mgr not available, waiting (5/15)... 2026-03-06T22:15:41.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:41 vm03 bash[17055]: audit 2026-03-06T21:15:41.357108+0000 mon.vm03 (mon.0) 17 : audit [DBG] from='client.? 192.168.123.103:0/3604218327' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T22:15:41.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:41 vm03 bash[17055]: audit 2026-03-06T21:15:41.357108+0000 mon.vm03 (mon.0) 17 : audit [DBG] from='client.? 192.168.123.103:0/3604218327' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: cluster 2026-03-06T21:15:42.358620+0000 mon.vm03 (mon.0) 18 : cluster [INF] Activating manager daemon vm03.uwuzgl 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: cluster 2026-03-06T21:15:42.358620+0000 mon.vm03 (mon.0) 18 : cluster [INF] Activating manager daemon vm03.uwuzgl 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: cluster 2026-03-06T21:15:42.363524+0000 mon.vm03 (mon.0) 19 : cluster [DBG] mgrmap e2: vm03.uwuzgl(active, starting, since 0.00503959s) 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: cluster 2026-03-06T21:15:42.363524+0000 mon.vm03 (mon.0) 19 : cluster [DBG] mgrmap e2: vm03.uwuzgl(active, starting, since 0.00503959s) 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.366071+0000 mon.vm03 (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.366071+0000 mon.vm03 (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.366409+0000 mon.vm03 (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.366409+0000 mon.vm03 (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.366715+0000 mon.vm03 (mon.0) 22 : audit [DBG] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.366715+0000 mon.vm03 (mon.0) 22 : audit [DBG] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.367005+0000 mon.vm03 (mon.0) 23 : audit [DBG] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.367005+0000 mon.vm03 (mon.0) 23 : audit [DBG] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.367296+0000 mon.vm03 (mon.0) 24 : audit [DBG] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr metadata", "who": "vm03.uwuzgl", "id": "vm03.uwuzgl"}]: dispatch 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.367296+0000 mon.vm03 (mon.0) 24 : audit [DBG] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr metadata", "who": "vm03.uwuzgl", "id": "vm03.uwuzgl"}]: dispatch 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: cluster 2026-03-06T21:15:42.372277+0000 mon.vm03 (mon.0) 25 : cluster [INF] Manager daemon vm03.uwuzgl is now available 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: cluster 2026-03-06T21:15:42.372277+0000 mon.vm03 (mon.0) 25 : cluster [INF] Manager daemon vm03.uwuzgl is now available 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.381214+0000 mon.vm03 (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/mirror_snapshot_schedule"}]: dispatch 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.381214+0000 mon.vm03 (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/mirror_snapshot_schedule"}]: dispatch 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.383619+0000 mon.vm03 (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.383619+0000 mon.vm03 (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.384045+0000 mon.vm03 (mon.0) 28 : audit [INF] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/trash_purge_schedule"}]: dispatch 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.384045+0000 mon.vm03 (mon.0) 28 : audit [INF] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/trash_purge_schedule"}]: dispatch 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.386242+0000 mon.vm03 (mon.0) 29 : audit [INF] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.386242+0000 mon.vm03 (mon.0) 29 : audit [INF] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.388502+0000 mon.vm03 (mon.0) 30 : audit [INF] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:42.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:42 vm03 bash[17055]: audit 2026-03-06T21:15:42.388502+0000 mon.vm03 (mon.0) 30 : audit [INF] from='mgr.14100 192.168.123.103:0/446384236' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsid": "894e000c-19a1-11f1-8dbe-23b24380a082", 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 0 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "vm03" 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_age": 12, 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T22:15:43.846 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T21:15:29:313048+0000", 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T21:15:29.313859+0000", 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-06T22:15:43.847 INFO:teuthology.orchestra.run.vm03.stdout:mgr is available 2026-03-06T22:15:44.212 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-06T22:15:44.212 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [global] 2026-03-06T22:15:44.212 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout fsid = 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:15:44.212 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-06T22:15:44.212 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.103:3300,v1:192.168.123.103:6789] 2026-03-06T22:15:44.212 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-06T22:15:44.212 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-06T22:15:44.212 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-06T22:15:44.212 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-06T22:15:44.213 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-06T22:15:44.213 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-06T22:15:44.213 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-06T22:15:44.213 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-06T22:15:44.213 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [osd] 2026-03-06T22:15:44.213 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-06T22:15:44.213 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-06T22:15:44.213 INFO:teuthology.orchestra.run.vm03.stdout:Enabling cephadm module... 2026-03-06T22:15:44.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:44 vm03 bash[17055]: cluster 2026-03-06T21:15:43.368836+0000 mon.vm03 (mon.0) 31 : cluster [DBG] mgrmap e3: vm03.uwuzgl(active, since 1.01035s) 2026-03-06T22:15:44.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:44 vm03 bash[17055]: cluster 2026-03-06T21:15:43.368836+0000 mon.vm03 (mon.0) 31 : cluster [DBG] mgrmap e3: vm03.uwuzgl(active, since 1.01035s) 2026-03-06T22:15:44.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:44 vm03 bash[17055]: audit 2026-03-06T21:15:43.797241+0000 mon.vm03 (mon.0) 32 : audit [DBG] from='client.? 192.168.123.103:0/3404582004' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T22:15:44.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:44 vm03 bash[17055]: audit 2026-03-06T21:15:43.797241+0000 mon.vm03 (mon.0) 32 : audit [DBG] from='client.? 192.168.123.103:0/3404582004' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T22:15:44.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:44 vm03 bash[17055]: audit 2026-03-06T21:15:44.160971+0000 mon.vm03 (mon.0) 33 : audit [INF] from='client.? 192.168.123.103:0/3713130269' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-06T22:15:44.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:44 vm03 bash[17055]: audit 2026-03-06T21:15:44.160971+0000 mon.vm03 (mon.0) 33 : audit [INF] from='client.? 192.168.123.103:0/3713130269' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-06T22:15:45.681 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:45 vm03 bash[17055]: cluster 2026-03-06T21:15:44.378478+0000 mon.vm03 (mon.0) 34 : cluster [DBG] mgrmap e4: vm03.uwuzgl(active, since 2s) 2026-03-06T22:15:45.682 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:45 vm03 bash[17055]: cluster 2026-03-06T21:15:44.378478+0000 mon.vm03 (mon.0) 34 : cluster [DBG] mgrmap e4: vm03.uwuzgl(active, since 2s) 2026-03-06T22:15:45.682 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:45 vm03 bash[17055]: audit 2026-03-06T21:15:44.573483+0000 mon.vm03 (mon.0) 35 : audit [INF] from='client.? 192.168.123.103:0/3053895857' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-06T22:15:45.682 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:45 vm03 bash[17055]: audit 2026-03-06T21:15:44.573483+0000 mon.vm03 (mon.0) 35 : audit [INF] from='client.? 192.168.123.103:0/3053895857' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-06T22:15:45.910 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-06T22:15:45.910 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 5, 2026-03-06T22:15:45.910 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-06T22:15:45.910 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "active_name": "vm03.uwuzgl", 2026-03-06T22:15:45.910 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-06T22:15:45.910 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-06T22:15:45.910 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for the mgr to restart... 2026-03-06T22:15:45.910 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mgr epoch 5... 2026-03-06T22:15:46.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:46 vm03 bash[17055]: audit 2026-03-06T21:15:45.379894+0000 mon.vm03 (mon.0) 36 : audit [INF] from='client.? 192.168.123.103:0/3053895857' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-06T22:15:46.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:46 vm03 bash[17055]: audit 2026-03-06T21:15:45.379894+0000 mon.vm03 (mon.0) 36 : audit [INF] from='client.? 192.168.123.103:0/3053895857' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-06T22:15:46.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:46 vm03 bash[17055]: cluster 2026-03-06T21:15:45.383051+0000 mon.vm03 (mon.0) 37 : cluster [DBG] mgrmap e5: vm03.uwuzgl(active, since 3s) 2026-03-06T22:15:46.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:46 vm03 bash[17055]: cluster 2026-03-06T21:15:45.383051+0000 mon.vm03 (mon.0) 37 : cluster [DBG] mgrmap e5: vm03.uwuzgl(active, since 3s) 2026-03-06T22:15:46.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:46 vm03 bash[17055]: audit 2026-03-06T21:15:45.844793+0000 mon.vm03 (mon.0) 38 : audit [DBG] from='client.? 192.168.123.103:0/3301488230' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-06T22:15:46.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:46 vm03 bash[17055]: audit 2026-03-06T21:15:45.844793+0000 mon.vm03 (mon.0) 38 : audit [DBG] from='client.? 192.168.123.103:0/3301488230' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-06T22:15:56.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:55 vm03 bash[17055]: cluster 2026-03-06T21:15:55.868954+0000 mon.vm03 (mon.0) 39 : cluster [INF] Active manager daemon vm03.uwuzgl restarted 2026-03-06T22:15:56.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:55 vm03 bash[17055]: cluster 2026-03-06T21:15:55.868954+0000 mon.vm03 (mon.0) 39 : cluster [INF] Active manager daemon vm03.uwuzgl restarted 2026-03-06T22:15:56.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:55 vm03 bash[17055]: cluster 2026-03-06T21:15:55.869368+0000 mon.vm03 (mon.0) 40 : cluster [INF] Activating manager daemon vm03.uwuzgl 2026-03-06T22:15:56.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:55 vm03 bash[17055]: cluster 2026-03-06T21:15:55.869368+0000 mon.vm03 (mon.0) 40 : cluster [INF] Activating manager daemon vm03.uwuzgl 2026-03-06T22:15:56.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:55 vm03 bash[17055]: cluster 2026-03-06T21:15:55.956910+0000 mon.vm03 (mon.0) 41 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-06T22:15:56.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:55 vm03 bash[17055]: cluster 2026-03-06T21:15:55.956910+0000 mon.vm03 (mon.0) 41 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-06T22:15:56.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:55 vm03 bash[17055]: cluster 2026-03-06T21:15:55.957109+0000 mon.vm03 (mon.0) 42 : cluster [DBG] mgrmap e6: vm03.uwuzgl(active, starting, since 0.0878232s) 2026-03-06T22:15:56.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:55 vm03 bash[17055]: cluster 2026-03-06T21:15:55.957109+0000 mon.vm03 (mon.0) 42 : cluster [DBG] mgrmap e6: vm03.uwuzgl(active, starting, since 0.0878232s) 2026-03-06T22:15:57.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-06T22:15:57.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 6, 2026-03-06T22:15:57.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-06T22:15:57.067 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-06T22:15:57.067 INFO:teuthology.orchestra.run.vm03.stdout:mgr epoch 5 is available 2026-03-06T22:15:57.067 INFO:teuthology.orchestra.run.vm03.stdout:Setting orchestrator backend to cephadm... 2026-03-06T22:15:57.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:56 vm03 bash[17055]: audit 2026-03-06T21:15:55.961452+0000 mon.vm03 (mon.0) 43 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:56 vm03 bash[17055]: audit 2026-03-06T21:15:55.961452+0000 mon.vm03 (mon.0) 43 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:56 vm03 bash[17055]: audit 2026-03-06T21:15:55.974582+0000 mon.vm03 (mon.0) 44 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr metadata", "who": "vm03.uwuzgl", "id": "vm03.uwuzgl"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:56 vm03 bash[17055]: audit 2026-03-06T21:15:55.974582+0000 mon.vm03 (mon.0) 44 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr metadata", "who": "vm03.uwuzgl", "id": "vm03.uwuzgl"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:56 vm03 bash[17055]: audit 2026-03-06T21:15:55.974726+0000 mon.vm03 (mon.0) 45 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:56 vm03 bash[17055]: audit 2026-03-06T21:15:55.974726+0000 mon.vm03 (mon.0) 45 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:56 vm03 bash[17055]: audit 2026-03-06T21:15:55.974817+0000 mon.vm03 (mon.0) 46 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:56 vm03 bash[17055]: audit 2026-03-06T21:15:55.974817+0000 mon.vm03 (mon.0) 46 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:56 vm03 bash[17055]: audit 2026-03-06T21:15:55.974903+0000 mon.vm03 (mon.0) 47 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:56 vm03 bash[17055]: audit 2026-03-06T21:15:55.974903+0000 mon.vm03 (mon.0) 47 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: cluster 2026-03-06T21:15:55.979721+0000 mon.vm03 (mon.0) 48 : cluster [INF] Manager daemon vm03.uwuzgl is now available 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: cluster 2026-03-06T21:15:55.979721+0000 mon.vm03 (mon.0) 48 : cluster [INF] Manager daemon vm03.uwuzgl is now available 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: cephadm 2026-03-06T21:15:55.992694+0000 mgr.vm03.uwuzgl (mgr.14124) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: cephadm 2026-03-06T21:15:55.992694+0000 mgr.vm03.uwuzgl (mgr.14124) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:55.995236+0000 mon.vm03 (mon.0) 49 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:55.995236+0000 mon.vm03 (mon.0) 49 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:55.999045+0000 mon.vm03 (mon.0) 50 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:55.999045+0000 mon.vm03 (mon.0) 50 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:56.012744+0000 mon.vm03 (mon.0) 51 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:56.012744+0000 mon.vm03 (mon.0) 51 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:56.019055+0000 mon.vm03 (mon.0) 52 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/mirror_snapshot_schedule"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:56.019055+0000 mon.vm03 (mon.0) 52 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/mirror_snapshot_schedule"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:56.019326+0000 mon.vm03 (mon.0) 53 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:56.019326+0000 mon.vm03 (mon.0) 53 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:56.020376+0000 mon.vm03 (mon.0) 54 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/trash_purge_schedule"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:56.020376+0000 mon.vm03 (mon.0) 54 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/trash_purge_schedule"}]: dispatch 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:56.355993+0000 mon.vm03 (mon.0) 55 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:56.355993+0000 mon.vm03 (mon.0) 55 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:56.358471+0000 mon.vm03 (mon.0) 56 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:57.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:56.358471+0000 mon.vm03 (mon.0) 56 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:57.894 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-06T22:15:57.894 INFO:teuthology.orchestra.run.vm03.stdout:Generating ssh key... 2026-03-06T22:15:58.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: cluster 2026-03-06T21:15:56.969957+0000 mon.vm03 (mon.0) 57 : cluster [DBG] mgrmap e7: vm03.uwuzgl(active, since 1.10067s) 2026-03-06T22:15:58.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: cluster 2026-03-06T21:15:56.969957+0000 mon.vm03 (mon.0) 57 : cluster [DBG] mgrmap e7: vm03.uwuzgl(active, since 1.10067s) 2026-03-06T22:15:58.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:56.970618+0000 mgr.vm03.uwuzgl (mgr.14124) 2 : audit [DBG] from='client.14128 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-06T22:15:58.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:56.970618+0000 mgr.vm03.uwuzgl (mgr.14124) 2 : audit [DBG] from='client.14128 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-06T22:15:58.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:56.977166+0000 mgr.vm03.uwuzgl (mgr.14124) 3 : audit [DBG] from='client.14128 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-06T22:15:58.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:56.977166+0000 mgr.vm03.uwuzgl (mgr.14124) 3 : audit [DBG] from='client.14128 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-06T22:15:58.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:57.437051+0000 mon.vm03 (mon.0) 58 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:58.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:57.437051+0000 mon.vm03 (mon.0) 58 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:58.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:57.440767+0000 mon.vm03 (mon.0) 59 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:15:58.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:57.440767+0000 mon.vm03 (mon.0) 59 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:15:58.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:57.899197+0000 mon.vm03 (mon.0) 60 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:15:58.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:57 vm03 bash[17055]: audit 2026-03-06T21:15:57.899197+0000 mon.vm03 (mon.0) 60 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:15:58.753 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCdY6Q1EToIq3mCYixKNuoi+bloeKOhC9rFW4hC+YnLuismoyKqHqQPPrGTT8umhhe93c4UXKjR/GcKE9ey/nXlznwMTzedy3tKn0NCJXxFGrUfaShTEav321oV5wUZwEMA6VG6NcGb43f5Nt4pZm1Krb20o7PuRVQHC/mwZuHpveocV/NaN1MHkgvjb6JSf6Ba4ecfe/yGYPTveQv7jNtY+Qhe7LS3D70qO/+8otZEvfV6870F3KJh8r+PcUSd7bZmS/6R1w8IMkK/bjzR1H4ss1rDQXQaxBrBOOtFLDHusY099hD8m/QHIAqFmCo7+zxX4P/2JkxWQCew6wRYXvHTtjmLe0VKMclZirIqeUYLX8Ni2QjggQzLLqn4SktI7IMMXFYMlSmHLUzlNEJYntCykHbq04H8aHLVAl3vQJSRQCmIU6VlMhX+InTrB0G7yW+77AR3NH0Bb3oa67Xojwd15BqSd9BDvdyX9K+Qeo1AbM/cG8gTbffdJSCE3jcti8M= ceph-894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:15:58.753 INFO:teuthology.orchestra.run.vm03.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-06T22:15:58.753 INFO:teuthology.orchestra.run.vm03.stdout:Adding key to root@localhost authorized_keys... 2026-03-06T22:15:58.753 INFO:teuthology.orchestra.run.vm03.stdout:Adding host vm03... 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: audit 2026-03-06T21:15:57.433569+0000 mgr.vm03.uwuzgl (mgr.14124) 4 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: audit 2026-03-06T21:15:57.433569+0000 mgr.vm03.uwuzgl (mgr.14124) 4 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: cephadm 2026-03-06T21:15:57.685854+0000 mgr.vm03.uwuzgl (mgr.14124) 5 : cephadm [INF] [06/Mar/2026:21:15:57] ENGINE Bus STARTING 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: cephadm 2026-03-06T21:15:57.685854+0000 mgr.vm03.uwuzgl (mgr.14124) 5 : cephadm [INF] [06/Mar/2026:21:15:57] ENGINE Bus STARTING 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: cephadm 2026-03-06T21:15:57.797730+0000 mgr.vm03.uwuzgl (mgr.14124) 6 : cephadm [INF] [06/Mar/2026:21:15:57] ENGINE Serving on https://192.168.123.103:7150 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: cephadm 2026-03-06T21:15:57.797730+0000 mgr.vm03.uwuzgl (mgr.14124) 6 : cephadm [INF] [06/Mar/2026:21:15:57] ENGINE Serving on https://192.168.123.103:7150 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: cephadm 2026-03-06T21:15:57.798249+0000 mgr.vm03.uwuzgl (mgr.14124) 7 : cephadm [INF] [06/Mar/2026:21:15:57] ENGINE Client ('192.168.123.103', 33796) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: cephadm 2026-03-06T21:15:57.798249+0000 mgr.vm03.uwuzgl (mgr.14124) 7 : cephadm [INF] [06/Mar/2026:21:15:57] ENGINE Client ('192.168.123.103', 33796) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: audit 2026-03-06T21:15:57.841010+0000 mgr.vm03.uwuzgl (mgr.14124) 8 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: audit 2026-03-06T21:15:57.841010+0000 mgr.vm03.uwuzgl (mgr.14124) 8 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: cephadm 2026-03-06T21:15:57.898583+0000 mgr.vm03.uwuzgl (mgr.14124) 9 : cephadm [INF] [06/Mar/2026:21:15:57] ENGINE Serving on http://192.168.123.103:8765 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: cephadm 2026-03-06T21:15:57.898583+0000 mgr.vm03.uwuzgl (mgr.14124) 9 : cephadm [INF] [06/Mar/2026:21:15:57] ENGINE Serving on http://192.168.123.103:8765 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: cephadm 2026-03-06T21:15:57.898620+0000 mgr.vm03.uwuzgl (mgr.14124) 10 : cephadm [INF] [06/Mar/2026:21:15:57] ENGINE Bus STARTED 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: cephadm 2026-03-06T21:15:57.898620+0000 mgr.vm03.uwuzgl (mgr.14124) 10 : cephadm [INF] [06/Mar/2026:21:15:57] ENGINE Bus STARTED 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: audit 2026-03-06T21:15:58.207598+0000 mgr.vm03.uwuzgl (mgr.14124) 11 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: audit 2026-03-06T21:15:58.207598+0000 mgr.vm03.uwuzgl (mgr.14124) 11 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: cephadm 2026-03-06T21:15:58.207827+0000 mgr.vm03.uwuzgl (mgr.14124) 12 : cephadm [INF] Generating ssh key... 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: cephadm 2026-03-06T21:15:58.207827+0000 mgr.vm03.uwuzgl (mgr.14124) 12 : cephadm [INF] Generating ssh key... 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: audit 2026-03-06T21:15:58.304085+0000 mon.vm03 (mon.0) 61 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: audit 2026-03-06T21:15:58.304085+0000 mon.vm03 (mon.0) 61 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: audit 2026-03-06T21:15:58.307374+0000 mon.vm03 (mon.0) 62 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: audit 2026-03-06T21:15:58.307374+0000 mon.vm03 (mon.0) 62 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: cluster 2026-03-06T21:15:58.442006+0000 mon.vm03 (mon.0) 63 : cluster [DBG] mgrmap e8: vm03.uwuzgl(active, since 2s) 2026-03-06T22:15:59.662 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:15:59 vm03 bash[17055]: cluster 2026-03-06T21:15:58.442006+0000 mon.vm03 (mon.0) 63 : cluster [DBG] mgrmap e8: vm03.uwuzgl(active, since 2s) 2026-03-06T22:16:00.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:00 vm03 bash[17055]: audit 2026-03-06T21:15:58.703255+0000 mgr.vm03.uwuzgl (mgr.14124) 13 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:00.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:00 vm03 bash[17055]: audit 2026-03-06T21:15:58.703255+0000 mgr.vm03.uwuzgl (mgr.14124) 13 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:00.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:00 vm03 bash[17055]: audit 2026-03-06T21:15:59.074997+0000 mgr.vm03.uwuzgl (mgr.14124) 14 : audit [DBG] from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm03", "addr": "192.168.123.103", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:00.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:00 vm03 bash[17055]: audit 2026-03-06T21:15:59.074997+0000 mgr.vm03.uwuzgl (mgr.14124) 14 : audit [DBG] from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm03", "addr": "192.168.123.103", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:01.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:01 vm03 bash[17055]: cephadm 2026-03-06T21:15:59.969404+0000 mgr.vm03.uwuzgl (mgr.14124) 15 : cephadm [INF] Deploying cephadm binary to vm03 2026-03-06T22:16:01.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:01 vm03 bash[17055]: cephadm 2026-03-06T21:15:59.969404+0000 mgr.vm03.uwuzgl (mgr.14124) 15 : cephadm [INF] Deploying cephadm binary to vm03 2026-03-06T22:16:01.880 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Added host 'vm03' with addr '192.168.123.103' 2026-03-06T22:16:01.880 INFO:teuthology.orchestra.run.vm03.stdout:Deploying mon service with default placement... 2026-03-06T22:16:02.608 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-06T22:16:02.608 INFO:teuthology.orchestra.run.vm03.stdout:Deploying mgr service with default placement... 2026-03-06T22:16:02.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:02 vm03 bash[17055]: audit 2026-03-06T21:16:01.814552+0000 mon.vm03 (mon.0) 64 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:02.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:02 vm03 bash[17055]: audit 2026-03-06T21:16:01.814552+0000 mon.vm03 (mon.0) 64 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:02.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:02 vm03 bash[17055]: cephadm 2026-03-06T21:16:01.814906+0000 mgr.vm03.uwuzgl (mgr.14124) 16 : cephadm [INF] Added host vm03 2026-03-06T22:16:02.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:02 vm03 bash[17055]: cephadm 2026-03-06T21:16:01.814906+0000 mgr.vm03.uwuzgl (mgr.14124) 16 : cephadm [INF] Added host vm03 2026-03-06T22:16:02.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:02 vm03 bash[17055]: audit 2026-03-06T21:16:01.815130+0000 mon.vm03 (mon.0) 65 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:16:02.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:02 vm03 bash[17055]: audit 2026-03-06T21:16:01.815130+0000 mon.vm03 (mon.0) 65 : audit [DBG] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:16:02.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:02 vm03 bash[17055]: audit 2026-03-06T21:16:02.554142+0000 mon.vm03 (mon.0) 66 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:02.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:02 vm03 bash[17055]: audit 2026-03-06T21:16:02.554142+0000 mon.vm03 (mon.0) 66 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:02.993 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-06T22:16:02.993 INFO:teuthology.orchestra.run.vm03.stdout:Deploying crash service with default placement... 2026-03-06T22:16:03.477 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Scheduled crash update... 2026-03-06T22:16:03.477 INFO:teuthology.orchestra.run.vm03.stdout:Deploying ceph-exporter service with default placement... 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: audit 2026-03-06T21:16:02.550189+0000 mgr.vm03.uwuzgl (mgr.14124) 17 : audit [DBG] from='client.14146 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: audit 2026-03-06T21:16:02.550189+0000 mgr.vm03.uwuzgl (mgr.14124) 17 : audit [DBG] from='client.14146 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: cephadm 2026-03-06T21:16:02.551066+0000 mgr.vm03.uwuzgl (mgr.14124) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: cephadm 2026-03-06T21:16:02.551066+0000 mgr.vm03.uwuzgl (mgr.14124) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: audit 2026-03-06T21:16:02.938417+0000 mgr.vm03.uwuzgl (mgr.14124) 19 : audit [DBG] from='client.14148 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: audit 2026-03-06T21:16:02.938417+0000 mgr.vm03.uwuzgl (mgr.14124) 19 : audit [DBG] from='client.14148 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: cephadm 2026-03-06T21:16:02.939067+0000 mgr.vm03.uwuzgl (mgr.14124) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: cephadm 2026-03-06T21:16:02.939067+0000 mgr.vm03.uwuzgl (mgr.14124) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: audit 2026-03-06T21:16:02.941615+0000 mon.vm03 (mon.0) 67 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: audit 2026-03-06T21:16:02.941615+0000 mon.vm03 (mon.0) 67 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: audit 2026-03-06T21:16:03.393208+0000 mgr.vm03.uwuzgl (mgr.14124) 21 : audit [DBG] from='client.14150 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "crash", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: audit 2026-03-06T21:16:03.393208+0000 mgr.vm03.uwuzgl (mgr.14124) 21 : audit [DBG] from='client.14150 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "crash", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: cephadm 2026-03-06T21:16:03.393939+0000 mgr.vm03.uwuzgl (mgr.14124) 22 : cephadm [INF] Saving service crash spec with placement * 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: cephadm 2026-03-06T21:16:03.393939+0000 mgr.vm03.uwuzgl (mgr.14124) 22 : cephadm [INF] Saving service crash spec with placement * 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: audit 2026-03-06T21:16:03.397076+0000 mon.vm03 (mon.0) 68 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: audit 2026-03-06T21:16:03.397076+0000 mon.vm03 (mon.0) 68 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: audit 2026-03-06T21:16:03.408273+0000 mon.vm03 (mon.0) 69 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: audit 2026-03-06T21:16:03.408273+0000 mon.vm03 (mon.0) 69 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:04.170 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: audit 2026-03-06T21:16:03.893131+0000 mon.vm03 (mon.0) 70 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:04.171 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: audit 2026-03-06T21:16:03.893131+0000 mon.vm03 (mon.0) 70 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:04.171 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: audit 2026-03-06T21:16:03.938522+0000 mon.vm03 (mon.0) 71 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:04.171 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:04 vm03 bash[17055]: audit 2026-03-06T21:16:03.938522+0000 mon.vm03 (mon.0) 71 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:04.184 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Scheduled ceph-exporter update... 2026-03-06T22:16:04.184 INFO:teuthology.orchestra.run.vm03.stdout:Deploying prometheus service with default placement... 2026-03-06T22:16:04.646 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Scheduled prometheus update... 2026-03-06T22:16:04.646 INFO:teuthology.orchestra.run.vm03.stdout:Deploying grafana service with default placement... 2026-03-06T22:16:05.051 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Scheduled grafana update... 2026-03-06T22:16:05.051 INFO:teuthology.orchestra.run.vm03.stdout:Deploying node-exporter service with default placement... 2026-03-06T22:16:05.396 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:05 vm03 bash[17055]: audit 2026-03-06T21:16:03.844509+0000 mgr.vm03.uwuzgl (mgr.14124) 23 : audit [DBG] from='client.14152 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "ceph-exporter", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:05.396 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:05 vm03 bash[17055]: audit 2026-03-06T21:16:03.844509+0000 mgr.vm03.uwuzgl (mgr.14124) 23 : audit [DBG] from='client.14152 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "ceph-exporter", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:05.396 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:05 vm03 bash[17055]: cephadm 2026-03-06T21:16:03.845287+0000 mgr.vm03.uwuzgl (mgr.14124) 24 : cephadm [INF] Saving service ceph-exporter spec with placement * 2026-03-06T22:16:05.396 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:05 vm03 bash[17055]: cephadm 2026-03-06T21:16:03.845287+0000 mgr.vm03.uwuzgl (mgr.14124) 24 : cephadm [INF] Saving service ceph-exporter spec with placement * 2026-03-06T22:16:05.396 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:05 vm03 bash[17055]: audit 2026-03-06T21:16:04.596632+0000 mon.vm03 (mon.0) 72 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:05.396 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:05 vm03 bash[17055]: audit 2026-03-06T21:16:04.596632+0000 mon.vm03 (mon.0) 72 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:05.396 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:05 vm03 bash[17055]: audit 2026-03-06T21:16:04.991080+0000 mon.vm03 (mon.0) 73 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:05.396 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:05 vm03 bash[17055]: audit 2026-03-06T21:16:04.991080+0000 mon.vm03 (mon.0) 73 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:05.434 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Scheduled node-exporter update... 2026-03-06T22:16:05.434 INFO:teuthology.orchestra.run.vm03.stdout:Deploying alertmanager service with default placement... 2026-03-06T22:16:05.819 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Scheduled alertmanager update... 2026-03-06T22:16:06.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: audit 2026-03-06T21:16:04.592232+0000 mgr.vm03.uwuzgl (mgr.14124) 25 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "prometheus", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:06.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: audit 2026-03-06T21:16:04.592232+0000 mgr.vm03.uwuzgl (mgr.14124) 25 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "prometheus", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:06.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: cephadm 2026-03-06T21:16:04.593175+0000 mgr.vm03.uwuzgl (mgr.14124) 26 : cephadm [INF] Saving service prometheus spec with placement count:1 2026-03-06T22:16:06.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: cephadm 2026-03-06T21:16:04.593175+0000 mgr.vm03.uwuzgl (mgr.14124) 26 : cephadm [INF] Saving service prometheus spec with placement count:1 2026-03-06T22:16:06.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: audit 2026-03-06T21:16:04.986668+0000 mgr.vm03.uwuzgl (mgr.14124) 27 : audit [DBG] from='client.14156 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "grafana", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:06.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: audit 2026-03-06T21:16:04.986668+0000 mgr.vm03.uwuzgl (mgr.14124) 27 : audit [DBG] from='client.14156 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "grafana", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:06.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: cephadm 2026-03-06T21:16:04.987726+0000 mgr.vm03.uwuzgl (mgr.14124) 28 : cephadm [INF] Saving service grafana spec with placement count:1 2026-03-06T22:16:06.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: cephadm 2026-03-06T21:16:04.987726+0000 mgr.vm03.uwuzgl (mgr.14124) 28 : cephadm [INF] Saving service grafana spec with placement count:1 2026-03-06T22:16:06.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: audit 2026-03-06T21:16:05.374821+0000 mgr.vm03.uwuzgl (mgr.14124) 29 : audit [DBG] from='client.14158 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "node-exporter", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:06.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: audit 2026-03-06T21:16:05.374821+0000 mgr.vm03.uwuzgl (mgr.14124) 29 : audit [DBG] from='client.14158 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "node-exporter", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:06.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: cephadm 2026-03-06T21:16:05.375466+0000 mgr.vm03.uwuzgl (mgr.14124) 30 : cephadm [INF] Saving service node-exporter spec with placement * 2026-03-06T22:16:06.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: cephadm 2026-03-06T21:16:05.375466+0000 mgr.vm03.uwuzgl (mgr.14124) 30 : cephadm [INF] Saving service node-exporter spec with placement * 2026-03-06T22:16:06.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: audit 2026-03-06T21:16:05.378777+0000 mon.vm03 (mon.0) 74 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:06.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: audit 2026-03-06T21:16:05.378777+0000 mon.vm03 (mon.0) 74 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:06.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: audit 2026-03-06T21:16:05.769600+0000 mon.vm03 (mon.0) 75 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:06.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: audit 2026-03-06T21:16:05.769600+0000 mon.vm03 (mon.0) 75 : audit [INF] from='mgr.14124 192.168.123.103:0/3629171052' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:06.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: audit 2026-03-06T21:16:06.150599+0000 mon.vm03 (mon.0) 76 : audit [INF] from='client.? 192.168.123.103:0/2959807028' entity='client.admin' 2026-03-06T22:16:06.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:06 vm03 bash[17055]: audit 2026-03-06T21:16:06.150599+0000 mon.vm03 (mon.0) 76 : audit [INF] from='client.? 192.168.123.103:0/2959807028' entity='client.admin' 2026-03-06T22:16:06.581 INFO:teuthology.orchestra.run.vm03.stdout:Enabling the dashboard module... 2026-03-06T22:16:07.641 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:07 vm03 bash[17055]: audit 2026-03-06T21:16:05.765697+0000 mgr.vm03.uwuzgl (mgr.14124) 31 : audit [DBG] from='client.14160 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "alertmanager", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:07.641 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:07 vm03 bash[17055]: audit 2026-03-06T21:16:05.765697+0000 mgr.vm03.uwuzgl (mgr.14124) 31 : audit [DBG] from='client.14160 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "alertmanager", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:07.641 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:07 vm03 bash[17055]: cephadm 2026-03-06T21:16:05.766407+0000 mgr.vm03.uwuzgl (mgr.14124) 32 : cephadm [INF] Saving service alertmanager spec with placement count:1 2026-03-06T22:16:07.641 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:07 vm03 bash[17055]: cephadm 2026-03-06T21:16:05.766407+0000 mgr.vm03.uwuzgl (mgr.14124) 32 : cephadm [INF] Saving service alertmanager spec with placement count:1 2026-03-06T22:16:07.641 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:07 vm03 bash[17055]: audit 2026-03-06T21:16:06.527574+0000 mon.vm03 (mon.0) 77 : audit [INF] from='client.? 192.168.123.103:0/851661528' entity='client.admin' 2026-03-06T22:16:07.641 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:07 vm03 bash[17055]: audit 2026-03-06T21:16:06.527574+0000 mon.vm03 (mon.0) 77 : audit [INF] from='client.? 192.168.123.103:0/851661528' entity='client.admin' 2026-03-06T22:16:07.641 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:07 vm03 bash[17055]: audit 2026-03-06T21:16:06.922054+0000 mon.vm03 (mon.0) 78 : audit [INF] from='client.? 192.168.123.103:0/193665833' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-06T22:16:07.641 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:07 vm03 bash[17055]: audit 2026-03-06T21:16:06.922054+0000 mon.vm03 (mon.0) 78 : audit [INF] from='client.? 192.168.123.103:0/193665833' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-06T22:16:08.047 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-06T22:16:08.047 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 9, 2026-03-06T22:16:08.047 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-06T22:16:08.047 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "active_name": "vm03.uwuzgl", 2026-03-06T22:16:08.047 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-06T22:16:08.047 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-06T22:16:08.047 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for the mgr to restart... 2026-03-06T22:16:08.047 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mgr epoch 9... 2026-03-06T22:16:08.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:08 vm03 bash[17055]: audit 2026-03-06T21:16:07.528932+0000 mon.vm03 (mon.0) 79 : audit [INF] from='client.? 192.168.123.103:0/193665833' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-06T22:16:08.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:08 vm03 bash[17055]: audit 2026-03-06T21:16:07.528932+0000 mon.vm03 (mon.0) 79 : audit [INF] from='client.? 192.168.123.103:0/193665833' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-06T22:16:08.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:08 vm03 bash[17055]: cluster 2026-03-06T21:16:07.531630+0000 mon.vm03 (mon.0) 80 : cluster [DBG] mgrmap e9: vm03.uwuzgl(active, since 11s) 2026-03-06T22:16:08.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:08 vm03 bash[17055]: cluster 2026-03-06T21:16:07.531630+0000 mon.vm03 (mon.0) 80 : cluster [DBG] mgrmap e9: vm03.uwuzgl(active, since 11s) 2026-03-06T22:16:08.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:08 vm03 bash[17055]: audit 2026-03-06T21:16:07.983752+0000 mon.vm03 (mon.0) 81 : audit [DBG] from='client.? 192.168.123.103:0/3786361270' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-06T22:16:08.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:08 vm03 bash[17055]: audit 2026-03-06T21:16:07.983752+0000 mon.vm03 (mon.0) 81 : audit [DBG] from='client.? 192.168.123.103:0/3786361270' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: cluster 2026-03-06T21:16:17.853980+0000 mon.vm03 (mon.0) 82 : cluster [INF] Active manager daemon vm03.uwuzgl restarted 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: cluster 2026-03-06T21:16:17.853980+0000 mon.vm03 (mon.0) 82 : cluster [INF] Active manager daemon vm03.uwuzgl restarted 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: cluster 2026-03-06T21:16:17.854176+0000 mon.vm03 (mon.0) 83 : cluster [INF] Activating manager daemon vm03.uwuzgl 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: cluster 2026-03-06T21:16:17.854176+0000 mon.vm03 (mon.0) 83 : cluster [INF] Activating manager daemon vm03.uwuzgl 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: cluster 2026-03-06T21:16:17.859119+0000 mon.vm03 (mon.0) 84 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: cluster 2026-03-06T21:16:17.859119+0000 mon.vm03 (mon.0) 84 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: cluster 2026-03-06T21:16:17.859217+0000 mon.vm03 (mon.0) 85 : cluster [DBG] mgrmap e10: vm03.uwuzgl(active, starting, since 0.00513879s) 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: cluster 2026-03-06T21:16:17.859217+0000 mon.vm03 (mon.0) 85 : cluster [DBG] mgrmap e10: vm03.uwuzgl(active, starting, since 0.00513879s) 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: audit 2026-03-06T21:16:17.862605+0000 mon.vm03 (mon.0) 86 : audit [DBG] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: audit 2026-03-06T21:16:17.862605+0000 mon.vm03 (mon.0) 86 : audit [DBG] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: audit 2026-03-06T21:16:17.862936+0000 mon.vm03 (mon.0) 87 : audit [DBG] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr metadata", "who": "vm03.uwuzgl", "id": "vm03.uwuzgl"}]: dispatch 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: audit 2026-03-06T21:16:17.862936+0000 mon.vm03 (mon.0) 87 : audit [DBG] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr metadata", "who": "vm03.uwuzgl", "id": "vm03.uwuzgl"}]: dispatch 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: audit 2026-03-06T21:16:17.863567+0000 mon.vm03 (mon.0) 88 : audit [DBG] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: audit 2026-03-06T21:16:17.863567+0000 mon.vm03 (mon.0) 88 : audit [DBG] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: audit 2026-03-06T21:16:17.863686+0000 mon.vm03 (mon.0) 89 : audit [DBG] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: audit 2026-03-06T21:16:17.863686+0000 mon.vm03 (mon.0) 89 : audit [DBG] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: audit 2026-03-06T21:16:17.863754+0000 mon.vm03 (mon.0) 90 : audit [DBG] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: audit 2026-03-06T21:16:17.863754+0000 mon.vm03 (mon.0) 90 : audit [DBG] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: cluster 2026-03-06T21:16:17.868859+0000 mon.vm03 (mon.0) 91 : cluster [INF] Manager daemon vm03.uwuzgl is now available 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: cluster 2026-03-06T21:16:17.868859+0000 mon.vm03 (mon.0) 91 : cluster [INF] Manager daemon vm03.uwuzgl is now available 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: audit 2026-03-06T21:16:17.892882+0000 mon.vm03 (mon.0) 92 : audit [DBG] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: audit 2026-03-06T21:16:17.892882+0000 mon.vm03 (mon.0) 92 : audit [DBG] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: audit 2026-03-06T21:16:17.902684+0000 mon.vm03 (mon.0) 93 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/mirror_snapshot_schedule"}]: dispatch 2026-03-06T22:16:18.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:17 vm03 bash[17055]: audit 2026-03-06T21:16:17.902684+0000 mon.vm03 (mon.0) 93 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/mirror_snapshot_schedule"}]: dispatch 2026-03-06T22:16:19.169 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-06T22:16:19.169 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 11, 2026-03-06T22:16:19.169 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-06T22:16:19.169 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-06T22:16:19.169 INFO:teuthology.orchestra.run.vm03.stdout:mgr epoch 9 is available 2026-03-06T22:16:19.169 INFO:teuthology.orchestra.run.vm03.stdout:Generating a dashboard self-signed certificate... 2026-03-06T22:16:19.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:19 vm03 bash[17055]: audit 2026-03-06T21:16:17.917110+0000 mon.vm03 (mon.0) 94 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/trash_purge_schedule"}]: dispatch 2026-03-06T22:16:19.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:19 vm03 bash[17055]: audit 2026-03-06T21:16:17.917110+0000 mon.vm03 (mon.0) 94 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/trash_purge_schedule"}]: dispatch 2026-03-06T22:16:20.174 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: cephadm 2026-03-06T21:16:18.868223+0000 mgr.vm03.uwuzgl (mgr.14168) 1 : cephadm [INF] [06/Mar/2026:21:16:18] ENGINE Bus STARTING 2026-03-06T22:16:20.174 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: cephadm 2026-03-06T21:16:18.868223+0000 mgr.vm03.uwuzgl (mgr.14168) 1 : cephadm [INF] [06/Mar/2026:21:16:18] ENGINE Bus STARTING 2026-03-06T22:16:20.174 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: cephadm 2026-03-06T21:16:18.979176+0000 mgr.vm03.uwuzgl (mgr.14168) 2 : cephadm [INF] [06/Mar/2026:21:16:18] ENGINE Serving on https://192.168.123.103:7150 2026-03-06T22:16:20.174 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: cephadm 2026-03-06T21:16:18.979176+0000 mgr.vm03.uwuzgl (mgr.14168) 2 : cephadm [INF] [06/Mar/2026:21:16:18] ENGINE Serving on https://192.168.123.103:7150 2026-03-06T22:16:20.174 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: cephadm 2026-03-06T21:16:18.979828+0000 mgr.vm03.uwuzgl (mgr.14168) 3 : cephadm [INF] [06/Mar/2026:21:16:18] ENGINE Client ('192.168.123.103', 39156) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T22:16:20.174 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: cephadm 2026-03-06T21:16:18.979828+0000 mgr.vm03.uwuzgl (mgr.14168) 3 : cephadm [INF] [06/Mar/2026:21:16:18] ENGINE Client ('192.168.123.103', 39156) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T22:16:20.174 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: audit 2026-03-06T21:16:19.042105+0000 mgr.vm03.uwuzgl (mgr.14168) 4 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-06T22:16:20.174 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: audit 2026-03-06T21:16:19.042105+0000 mgr.vm03.uwuzgl (mgr.14168) 4 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-06T22:16:20.174 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: cluster 2026-03-06T21:16:19.043336+0000 mon.vm03 (mon.0) 95 : cluster [DBG] mgrmap e11: vm03.uwuzgl(active, since 1.18926s) 2026-03-06T22:16:20.174 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: cluster 2026-03-06T21:16:19.043336+0000 mon.vm03 (mon.0) 95 : cluster [DBG] mgrmap e11: vm03.uwuzgl(active, since 1.18926s) 2026-03-06T22:16:20.174 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: audit 2026-03-06T21:16:19.046871+0000 mgr.vm03.uwuzgl (mgr.14168) 5 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-06T22:16:20.174 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: audit 2026-03-06T21:16:19.046871+0000 mgr.vm03.uwuzgl (mgr.14168) 5 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-06T22:16:20.175 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: cephadm 2026-03-06T21:16:19.080812+0000 mgr.vm03.uwuzgl (mgr.14168) 6 : cephadm [INF] [06/Mar/2026:21:16:19] ENGINE Serving on http://192.168.123.103:8765 2026-03-06T22:16:20.175 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: cephadm 2026-03-06T21:16:19.080812+0000 mgr.vm03.uwuzgl (mgr.14168) 6 : cephadm [INF] [06/Mar/2026:21:16:19] ENGINE Serving on http://192.168.123.103:8765 2026-03-06T22:16:20.175 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: cephadm 2026-03-06T21:16:19.080863+0000 mgr.vm03.uwuzgl (mgr.14168) 7 : cephadm [INF] [06/Mar/2026:21:16:19] ENGINE Bus STARTED 2026-03-06T22:16:20.175 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: cephadm 2026-03-06T21:16:19.080863+0000 mgr.vm03.uwuzgl (mgr.14168) 7 : cephadm [INF] [06/Mar/2026:21:16:19] ENGINE Bus STARTED 2026-03-06T22:16:20.175 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: audit 2026-03-06T21:16:19.500218+0000 mgr.vm03.uwuzgl (mgr.14168) 8 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:20.175 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: audit 2026-03-06T21:16:19.500218+0000 mgr.vm03.uwuzgl (mgr.14168) 8 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:20.175 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: audit 2026-03-06T21:16:19.684380+0000 mon.vm03 (mon.0) 96 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:20.175 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: audit 2026-03-06T21:16:19.684380+0000 mon.vm03 (mon.0) 96 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:20.175 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: audit 2026-03-06T21:16:19.894587+0000 mon.vm03 (mon.0) 97 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:20.175 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:20 vm03 bash[17055]: audit 2026-03-06T21:16:19.894587+0000 mon.vm03 (mon.0) 97 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:20.560 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-06T22:16:20.560 INFO:teuthology.orchestra.run.vm03.stdout:Creating initial admin user... 2026-03-06T22:16:21.204 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$1Lw6MVL2R8iS0fYfOcreG.baD017DZRP/F5orqTW0jhgDqB87Kwg6", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1772831781, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-06T22:16:21.204 INFO:teuthology.orchestra.run.vm03.stdout:Fetching dashboard port number... 2026-03-06T22:16:21.572 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 8443 2026-03-06T22:16:21.572 INFO:teuthology.orchestra.run.vm03.stdout:firewalld does not appear to be present 2026-03-06T22:16:21.572 INFO:teuthology.orchestra.run.vm03.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-06T22:16:21.574 INFO:teuthology.orchestra.run.vm03.stdout:Ceph Dashboard is now available at: 2026-03-06T22:16:21.574 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:16:21.574 INFO:teuthology.orchestra.run.vm03.stdout: URL: https://vm03.local:8443/ 2026-03-06T22:16:21.574 INFO:teuthology.orchestra.run.vm03.stdout: User: admin 2026-03-06T22:16:21.574 INFO:teuthology.orchestra.run.vm03.stdout: Password: cxsf4xw0iz 2026-03-06T22:16:21.574 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:16:21.574 INFO:teuthology.orchestra.run.vm03.stdout:Saving cluster configuration to /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config directory 2026-03-06T22:16:22.135 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:21 vm03 bash[17055]: cluster 2026-03-06T21:16:20.688433+0000 mon.vm03 (mon.0) 98 : cluster [DBG] mgrmap e12: vm03.uwuzgl(active, since 2s) 2026-03-06T22:16:22.135 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:21 vm03 bash[17055]: cluster 2026-03-06T21:16:20.688433+0000 mon.vm03 (mon.0) 98 : cluster [DBG] mgrmap e12: vm03.uwuzgl(active, since 2s) 2026-03-06T22:16:22.135 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:21 vm03 bash[17055]: audit 2026-03-06T21:16:20.994949+0000 mgr.vm03.uwuzgl (mgr.14168) 9 : audit [DBG] from='client.14182 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:22.135 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:21 vm03 bash[17055]: audit 2026-03-06T21:16:20.994949+0000 mgr.vm03.uwuzgl (mgr.14168) 9 : audit [DBG] from='client.14182 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:22.135 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:21 vm03 bash[17055]: audit 2026-03-06T21:16:21.148920+0000 mon.vm03 (mon.0) 99 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:22.135 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:21 vm03 bash[17055]: audit 2026-03-06T21:16:21.148920+0000 mon.vm03 (mon.0) 99 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:22.135 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:21 vm03 bash[17055]: audit 2026-03-06T21:16:21.514469+0000 mon.vm03 (mon.0) 100 : audit [DBG] from='client.? 192.168.123.103:0/4186722353' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-06T22:16:22.135 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:21 vm03 bash[17055]: audit 2026-03-06T21:16:21.514469+0000 mon.vm03 (mon.0) 100 : audit [DBG] from='client.? 192.168.123.103:0/4186722353' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout:Or, if you are only running a single cluster on this host: 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout: ceph telemetry on 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout:For more information see: 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:16:22.262 INFO:teuthology.orchestra.run.vm03.stdout:Bootstrap complete. 2026-03-06T22:16:22.280 INFO:tasks.cephadm:Fetching config... 2026-03-06T22:16:22.280 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-06T22:16:22.280 DEBUG:teuthology.orchestra.run.vm03:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-06T22:16:22.283 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-06T22:16:22.283 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-06T22:16:22.283 DEBUG:teuthology.orchestra.run.vm03:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-06T22:16:22.328 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-06T22:16:22.328 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-06T22:16:22.328 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/keyring of=/dev/stdout 2026-03-06T22:16:22.376 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-06T22:16:22.376 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-06T22:16:22.376 DEBUG:teuthology.orchestra.run.vm03:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-06T22:16:22.420 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-06T22:16:22.420 DEBUG:teuthology.orchestra.run.vm03:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCdY6Q1EToIq3mCYixKNuoi+bloeKOhC9rFW4hC+YnLuismoyKqHqQPPrGTT8umhhe93c4UXKjR/GcKE9ey/nXlznwMTzedy3tKn0NCJXxFGrUfaShTEav321oV5wUZwEMA6VG6NcGb43f5Nt4pZm1Krb20o7PuRVQHC/mwZuHpveocV/NaN1MHkgvjb6JSf6Ba4ecfe/yGYPTveQv7jNtY+Qhe7LS3D70qO/+8otZEvfV6870F3KJh8r+PcUSd7bZmS/6R1w8IMkK/bjzR1H4ss1rDQXQaxBrBOOtFLDHusY099hD8m/QHIAqFmCo7+zxX4P/2JkxWQCew6wRYXvHTtjmLe0VKMclZirIqeUYLX8Ni2QjggQzLLqn4SktI7IMMXFYMlSmHLUzlNEJYntCykHbq04H8aHLVAl3vQJSRQCmIU6VlMhX+InTrB0G7yW+77AR3NH0Bb3oa67Xojwd15BqSd9BDvdyX9K+Qeo1AbM/cG8gTbffdJSCE3jcti8M= ceph-894e000c-19a1-11f1-8dbe-23b24380a082' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-06T22:16:22.472 INFO:teuthology.orchestra.run.vm03.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCdY6Q1EToIq3mCYixKNuoi+bloeKOhC9rFW4hC+YnLuismoyKqHqQPPrGTT8umhhe93c4UXKjR/GcKE9ey/nXlznwMTzedy3tKn0NCJXxFGrUfaShTEav321oV5wUZwEMA6VG6NcGb43f5Nt4pZm1Krb20o7PuRVQHC/mwZuHpveocV/NaN1MHkgvjb6JSf6Ba4ecfe/yGYPTveQv7jNtY+Qhe7LS3D70qO/+8otZEvfV6870F3KJh8r+PcUSd7bZmS/6R1w8IMkK/bjzR1H4ss1rDQXQaxBrBOOtFLDHusY099hD8m/QHIAqFmCo7+zxX4P/2JkxWQCew6wRYXvHTtjmLe0VKMclZirIqeUYLX8Ni2QjggQzLLqn4SktI7IMMXFYMlSmHLUzlNEJYntCykHbq04H8aHLVAl3vQJSRQCmIU6VlMhX+InTrB0G7yW+77AR3NH0Bb3oa67Xojwd15BqSd9BDvdyX9K+Qeo1AbM/cG8gTbffdJSCE3jcti8M= ceph-894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:16:22.476 DEBUG:teuthology.orchestra.run.vm08:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCdY6Q1EToIq3mCYixKNuoi+bloeKOhC9rFW4hC+YnLuismoyKqHqQPPrGTT8umhhe93c4UXKjR/GcKE9ey/nXlznwMTzedy3tKn0NCJXxFGrUfaShTEav321oV5wUZwEMA6VG6NcGb43f5Nt4pZm1Krb20o7PuRVQHC/mwZuHpveocV/NaN1MHkgvjb6JSf6Ba4ecfe/yGYPTveQv7jNtY+Qhe7LS3D70qO/+8otZEvfV6870F3KJh8r+PcUSd7bZmS/6R1w8IMkK/bjzR1H4ss1rDQXQaxBrBOOtFLDHusY099hD8m/QHIAqFmCo7+zxX4P/2JkxWQCew6wRYXvHTtjmLe0VKMclZirIqeUYLX8Ni2QjggQzLLqn4SktI7IMMXFYMlSmHLUzlNEJYntCykHbq04H8aHLVAl3vQJSRQCmIU6VlMhX+InTrB0G7yW+77AR3NH0Bb3oa67Xojwd15BqSd9BDvdyX9K+Qeo1AbM/cG8gTbffdJSCE3jcti8M= ceph-894e000c-19a1-11f1-8dbe-23b24380a082' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-06T22:16:22.487 INFO:teuthology.orchestra.run.vm08.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCdY6Q1EToIq3mCYixKNuoi+bloeKOhC9rFW4hC+YnLuismoyKqHqQPPrGTT8umhhe93c4UXKjR/GcKE9ey/nXlznwMTzedy3tKn0NCJXxFGrUfaShTEav321oV5wUZwEMA6VG6NcGb43f5Nt4pZm1Krb20o7PuRVQHC/mwZuHpveocV/NaN1MHkgvjb6JSf6Ba4ecfe/yGYPTveQv7jNtY+Qhe7LS3D70qO/+8otZEvfV6870F3KJh8r+PcUSd7bZmS/6R1w8IMkK/bjzR1H4ss1rDQXQaxBrBOOtFLDHusY099hD8m/QHIAqFmCo7+zxX4P/2JkxWQCew6wRYXvHTtjmLe0VKMclZirIqeUYLX8Ni2QjggQzLLqn4SktI7IMMXFYMlSmHLUzlNEJYntCykHbq04H8aHLVAl3vQJSRQCmIU6VlMhX+InTrB0G7yW+77AR3NH0Bb3oa67Xojwd15BqSd9BDvdyX9K+Qeo1AbM/cG8gTbffdJSCE3jcti8M= ceph-894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:16:22.491 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-06T22:16:23.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:23 vm03 bash[17055]: audit 2026-03-06T21:16:22.197067+0000 mon.vm03 (mon.0) 101 : audit [INF] from='client.? 192.168.123.103:0/3276835725' entity='client.admin' 2026-03-06T22:16:23.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:23 vm03 bash[17055]: audit 2026-03-06T21:16:22.197067+0000 mon.vm03 (mon.0) 101 : audit [INF] from='client.? 192.168.123.103:0/3276835725' entity='client.admin' 2026-03-06T22:16:24.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:24 vm03 bash[17055]: audit 2026-03-06T21:16:23.226109+0000 mon.vm03 (mon.0) 102 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:24.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:24 vm03 bash[17055]: audit 2026-03-06T21:16:23.226109+0000 mon.vm03 (mon.0) 102 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:24.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:24 vm03 bash[17055]: audit 2026-03-06T21:16:23.814066+0000 mon.vm03 (mon.0) 103 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:24.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:24 vm03 bash[17055]: audit 2026-03-06T21:16:23.814066+0000 mon.vm03 (mon.0) 103 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:26.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:25 vm03 bash[17055]: cluster 2026-03-06T21:16:24.939622+0000 mon.vm03 (mon.0) 104 : cluster [DBG] mgrmap e13: vm03.uwuzgl(active, since 7s) 2026-03-06T22:16:26.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:25 vm03 bash[17055]: cluster 2026-03-06T21:16:24.939622+0000 mon.vm03 (mon.0) 104 : cluster [DBG] mgrmap e13: vm03.uwuzgl(active, since 7s) 2026-03-06T22:16:27.091 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:16:27.662 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-06T22:16:27.662 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-06T22:16:28.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:28 vm03 bash[17055]: audit 2026-03-06T21:16:27.507734+0000 mon.vm03 (mon.0) 105 : audit [INF] from='client.? 192.168.123.103:0/1036649459' entity='client.admin' 2026-03-06T22:16:28.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:28 vm03 bash[17055]: audit 2026-03-06T21:16:27.507734+0000 mon.vm03 (mon.0) 105 : audit [INF] from='client.? 192.168.123.103:0/1036649459' entity='client.admin' 2026-03-06T22:16:30.889 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:30 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:16:30.889 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:30 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:16:31.163 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.020508+0000 mon.vm03 (mon.0) 106 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:31.163 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.020508+0000 mon.vm03 (mon.0) 106 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:31.163 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.023617+0000 mon.vm03 (mon.0) 107 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:31.163 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.023617+0000 mon.vm03 (mon.0) 107 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:31.163 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.024651+0000 mon.vm03 (mon.0) 108 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:16:31.163 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.024651+0000 mon.vm03 (mon.0) 108 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:16:31.163 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.027733+0000 mon.vm03 (mon.0) 109 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:31.163 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.027733+0000 mon.vm03 (mon.0) 109 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:31.163 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.028839+0000 mon.vm03 (mon.0) 110 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm03", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T22:16:31.163 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.028839+0000 mon.vm03 (mon.0) 110 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm03", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T22:16:31.163 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.029840+0000 mon.vm03 (mon.0) 111 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm03", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-06T22:16:31.163 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.029840+0000 mon.vm03 (mon.0) 111 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm03", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-06T22:16:31.163 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.031180+0000 mon.vm03 (mon.0) 112 : audit [DBG] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:16:31.163 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.031180+0000 mon.vm03 (mon.0) 112 : audit [DBG] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:16:31.164 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: cephadm 2026-03-06T21:16:30.031796+0000 mgr.vm03.uwuzgl (mgr.14168) 10 : cephadm [INF] Deploying daemon ceph-exporter.vm03 on vm03 2026-03-06T22:16:31.164 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: cephadm 2026-03-06T21:16:30.031796+0000 mgr.vm03.uwuzgl (mgr.14168) 10 : cephadm [INF] Deploying daemon ceph-exporter.vm03 on vm03 2026-03-06T22:16:31.164 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.812127+0000 mon.vm03 (mon.0) 113 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:31.164 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.812127+0000 mon.vm03 (mon.0) 113 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:31.164 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.814412+0000 mon.vm03 (mon.0) 114 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:31.164 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.814412+0000 mon.vm03 (mon.0) 114 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:31.164 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.817601+0000 mon.vm03 (mon.0) 115 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:31.164 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.817601+0000 mon.vm03 (mon.0) 115 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:31.164 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.820809+0000 mon.vm03 (mon.0) 116 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:31.164 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.820809+0000 mon.vm03 (mon.0) 116 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:31.164 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.821862+0000 mon.vm03 (mon.0) 117 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm03", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T22:16:31.164 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.821862+0000 mon.vm03 (mon.0) 117 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm03", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T22:16:31.164 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.824010+0000 mon.vm03 (mon.0) 118 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm03", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-06T22:16:31.164 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.824010+0000 mon.vm03 (mon.0) 118 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm03", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-06T22:16:31.164 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.825194+0000 mon.vm03 (mon.0) 119 : audit [DBG] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:16:31.164 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 bash[17055]: audit 2026-03-06T21:16:30.825194+0000 mon.vm03 (mon.0) 119 : audit [DBG] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:16:31.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:16:31.882 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:31 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:16:32.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:32 vm03 bash[17055]: cephadm 2026-03-06T21:16:30.826247+0000 mgr.vm03.uwuzgl (mgr.14168) 11 : cephadm [INF] Deploying daemon crash.vm03 on vm03 2026-03-06T22:16:32.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:32 vm03 bash[17055]: cephadm 2026-03-06T21:16:30.826247+0000 mgr.vm03.uwuzgl (mgr.14168) 11 : cephadm [INF] Deploying daemon crash.vm03 on vm03 2026-03-06T22:16:32.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:32 vm03 bash[17055]: audit 2026-03-06T21:16:31.608598+0000 mon.vm03 (mon.0) 120 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:32.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:32 vm03 bash[17055]: audit 2026-03-06T21:16:31.608598+0000 mon.vm03 (mon.0) 120 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:32.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:32 vm03 bash[17055]: audit 2026-03-06T21:16:31.611056+0000 mon.vm03 (mon.0) 121 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:32.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:32 vm03 bash[17055]: audit 2026-03-06T21:16:31.611056+0000 mon.vm03 (mon.0) 121 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:32.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:32 vm03 bash[17055]: audit 2026-03-06T21:16:31.613287+0000 mon.vm03 (mon.0) 122 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:32.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:32 vm03 bash[17055]: audit 2026-03-06T21:16:31.613287+0000 mon.vm03 (mon.0) 122 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:32.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:32 vm03 bash[17055]: audit 2026-03-06T21:16:31.616294+0000 mon.vm03 (mon.0) 123 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:32.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:32 vm03 bash[17055]: audit 2026-03-06T21:16:31.616294+0000 mon.vm03 (mon.0) 123 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:32.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:32 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:16:32.584 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:32 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:16:33.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:33 vm03 bash[17055]: cephadm 2026-03-06T21:16:31.617496+0000 mgr.vm03.uwuzgl (mgr.14168) 12 : cephadm [INF] Deploying daemon node-exporter.vm03 on vm03 2026-03-06T22:16:33.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:33 vm03 bash[17055]: cephadm 2026-03-06T21:16:31.617496+0000 mgr.vm03.uwuzgl (mgr.14168) 12 : cephadm [INF] Deploying daemon node-exporter.vm03 on vm03 2026-03-06T22:16:33.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:33 vm03 bash[17055]: audit 2026-03-06T21:16:32.331917+0000 mon.vm03 (mon.0) 124 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:33.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:33 vm03 bash[17055]: audit 2026-03-06T21:16:32.331917+0000 mon.vm03 (mon.0) 124 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:33.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:33 vm03 bash[17055]: audit 2026-03-06T21:16:32.334247+0000 mon.vm03 (mon.0) 125 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:33.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:33 vm03 bash[17055]: audit 2026-03-06T21:16:32.334247+0000 mon.vm03 (mon.0) 125 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:33.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:33 vm03 bash[17055]: audit 2026-03-06T21:16:32.336392+0000 mon.vm03 (mon.0) 126 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:33.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:33 vm03 bash[17055]: audit 2026-03-06T21:16:32.336392+0000 mon.vm03 (mon.0) 126 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:33.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:33 vm03 bash[17055]: audit 2026-03-06T21:16:32.338170+0000 mon.vm03 (mon.0) 127 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:33.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:33 vm03 bash[17055]: audit 2026-03-06T21:16:32.338170+0000 mon.vm03 (mon.0) 127 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:33.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:33 vm03 bash[17055]: cephadm 2026-03-06T21:16:32.342486+0000 mgr.vm03.uwuzgl (mgr.14168) 13 : cephadm [INF] Deploying daemon alertmanager.vm03 on vm03 2026-03-06T22:16:33.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:33 vm03 bash[17055]: cephadm 2026-03-06T21:16:32.342486+0000 mgr.vm03.uwuzgl (mgr.14168) 13 : cephadm [INF] Deploying daemon alertmanager.vm03 on vm03 2026-03-06T22:16:33.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:33 vm03 bash[17055]: audit 2026-03-06T21:16:32.890570+0000 mon.vm03 (mon.0) 128 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:33.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:33 vm03 bash[17055]: audit 2026-03-06T21:16:32.890570+0000 mon.vm03 (mon.0) 128 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:34.158 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:16:34.600 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm08 2026-03-06T22:16:34.600 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T22:16:34.600 DEBUG:teuthology.orchestra.run.vm08:> dd of=/etc/ceph/ceph.conf 2026-03-06T22:16:34.604 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T22:16:34.604 DEBUG:teuthology.orchestra.run.vm08:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T22:16:34.647 INFO:tasks.cephadm:Adding host vm08 to orchestrator... 2026-03-06T22:16:34.647 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph orch host add vm08 2026-03-06T22:16:35.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:35 vm03 bash[17055]: audit 2026-03-06T21:16:34.527075+0000 mgr.vm03.uwuzgl (mgr.14168) 14 : audit [DBG] from='client.14191 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:35.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:35 vm03 bash[17055]: audit 2026-03-06T21:16:34.527075+0000 mgr.vm03.uwuzgl (mgr.14168) 14 : audit [DBG] from='client.14191 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:35.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:35 vm03 bash[17055]: audit 2026-03-06T21:16:34.530165+0000 mon.vm03 (mon.0) 129 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:35.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:35 vm03 bash[17055]: audit 2026-03-06T21:16:34.530165+0000 mon.vm03 (mon.0) 129 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:36.804 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:36 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:16:37.141 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:36 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.015730+0000 mon.vm03 (mon.0) 130 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.015730+0000 mon.vm03 (mon.0) 130 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.019521+0000 mon.vm03 (mon.0) 131 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.019521+0000 mon.vm03 (mon.0) 131 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.022660+0000 mon.vm03 (mon.0) 132 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.022660+0000 mon.vm03 (mon.0) 132 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.025419+0000 mon.vm03 (mon.0) 133 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.025419+0000 mon.vm03 (mon.0) 133 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.029106+0000 mon.vm03 (mon.0) 134 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.029106+0000 mon.vm03 (mon.0) 134 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.031909+0000 mon.vm03 (mon.0) 135 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.031909+0000 mon.vm03 (mon.0) 135 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: cephadm 2026-03-06T21:16:37.037554+0000 mgr.vm03.uwuzgl (mgr.14168) 15 : cephadm [INF] Regenerating cephadm self-signed grafana TLS certificates 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: cephadm 2026-03-06T21:16:37.037554+0000 mgr.vm03.uwuzgl (mgr.14168) 15 : cephadm [INF] Regenerating cephadm self-signed grafana TLS certificates 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.082935+0000 mon.vm03 (mon.0) 136 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.082935+0000 mon.vm03 (mon.0) 136 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.088277+0000 mon.vm03 (mon.0) 137 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.088277+0000 mon.vm03 (mon.0) 137 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.090352+0000 mon.vm03 (mon.0) 138 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.090352+0000 mon.vm03 (mon.0) 138 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.090724+0000 mgr.vm03.uwuzgl (mgr.14168) 16 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-06T22:16:38.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.090724+0000 mgr.vm03.uwuzgl (mgr.14168) 16 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-06T22:16:38.442 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.095532+0000 mon.vm03 (mon.0) 139 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.442 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.095532+0000 mon.vm03 (mon.0) 139 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.442 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: cephadm 2026-03-06T21:16:37.105073+0000 mgr.vm03.uwuzgl (mgr.14168) 17 : cephadm [INF] Deploying daemon grafana.vm03 on vm03 2026-03-06T22:16:38.442 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: cephadm 2026-03-06T21:16:37.105073+0000 mgr.vm03.uwuzgl (mgr.14168) 17 : cephadm [INF] Deploying daemon grafana.vm03 on vm03 2026-03-06T22:16:38.442 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.894514+0000 mon.vm03 (mon.0) 140 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:38.442 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:38 vm03 bash[17055]: audit 2026-03-06T21:16:37.894514+0000 mon.vm03 (mon.0) 140 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:39.436 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:16:39.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:39 vm03 bash[17055]: cluster 2026-03-06T21:16:37.864622+0000 mgr.vm03.uwuzgl (mgr.14168) 18 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:39.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:39 vm03 bash[17055]: cluster 2026-03-06T21:16:37.864622+0000 mgr.vm03.uwuzgl (mgr.14168) 18 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:41.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:41 vm03 bash[17055]: audit 2026-03-06T21:16:39.853661+0000 mgr.vm03.uwuzgl (mgr.14168) 19 : audit [DBG] from='client.14193 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm08", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:41.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:41 vm03 bash[17055]: audit 2026-03-06T21:16:39.853661+0000 mgr.vm03.uwuzgl (mgr.14168) 19 : audit [DBG] from='client.14193 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm08", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:16:41.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:41 vm03 bash[17055]: cluster 2026-03-06T21:16:39.864852+0000 mgr.vm03.uwuzgl (mgr.14168) 20 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:41.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:41 vm03 bash[17055]: cluster 2026-03-06T21:16:39.864852+0000 mgr.vm03.uwuzgl (mgr.14168) 20 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:42.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:42 vm03 bash[17055]: cephadm 2026-03-06T21:16:40.777418+0000 mgr.vm03.uwuzgl (mgr.14168) 21 : cephadm [INF] Deploying cephadm binary to vm08 2026-03-06T22:16:42.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:42 vm03 bash[17055]: cephadm 2026-03-06T21:16:40.777418+0000 mgr.vm03.uwuzgl (mgr.14168) 21 : cephadm [INF] Deploying cephadm binary to vm08 2026-03-06T22:16:42.684 INFO:teuthology.orchestra.run.vm03.stdout:Added host 'vm08' with addr '192.168.123.108' 2026-03-06T22:16:43.175 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph orch host ls --format=json 2026-03-06T22:16:43.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:43 vm03 bash[17055]: cluster 2026-03-06T21:16:41.865085+0000 mgr.vm03.uwuzgl (mgr.14168) 22 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:43.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:43 vm03 bash[17055]: cluster 2026-03-06T21:16:41.865085+0000 mgr.vm03.uwuzgl (mgr.14168) 22 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:43.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:43 vm03 bash[17055]: audit 2026-03-06T21:16:42.675854+0000 mon.vm03 (mon.0) 141 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:43.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:43 vm03 bash[17055]: audit 2026-03-06T21:16:42.675854+0000 mon.vm03 (mon.0) 141 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:44.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:44 vm03 bash[17055]: cephadm 2026-03-06T21:16:42.676247+0000 mgr.vm03.uwuzgl (mgr.14168) 23 : cephadm [INF] Added host vm08 2026-03-06T22:16:44.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:44 vm03 bash[17055]: cephadm 2026-03-06T21:16:42.676247+0000 mgr.vm03.uwuzgl (mgr.14168) 23 : cephadm [INF] Added host vm08 2026-03-06T22:16:45.197 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:45 vm03 bash[17055]: cluster 2026-03-06T21:16:43.865322+0000 mgr.vm03.uwuzgl (mgr.14168) 24 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:45.198 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:45 vm03 bash[17055]: cluster 2026-03-06T21:16:43.865322+0000 mgr.vm03.uwuzgl (mgr.14168) 24 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:46.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:46 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:16:46.557 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:46 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: cluster 2026-03-06T21:16:45.865578+0000 mgr.vm03.uwuzgl (mgr.14168) 25 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: cluster 2026-03-06T21:16:45.865578+0000 mgr.vm03.uwuzgl (mgr.14168) 25 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: audit 2026-03-06T21:16:46.373969+0000 mon.vm03 (mon.0) 142 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: audit 2026-03-06T21:16:46.373969+0000 mon.vm03 (mon.0) 142 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: audit 2026-03-06T21:16:46.378374+0000 mon.vm03 (mon.0) 143 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: audit 2026-03-06T21:16:46.378374+0000 mon.vm03 (mon.0) 143 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: audit 2026-03-06T21:16:46.381522+0000 mon.vm03 (mon.0) 144 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: audit 2026-03-06T21:16:46.381522+0000 mon.vm03 (mon.0) 144 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: audit 2026-03-06T21:16:46.384943+0000 mon.vm03 (mon.0) 145 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: audit 2026-03-06T21:16:46.384943+0000 mon.vm03 (mon.0) 145 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: audit 2026-03-06T21:16:46.387818+0000 mon.vm03 (mon.0) 146 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: audit 2026-03-06T21:16:46.387818+0000 mon.vm03 (mon.0) 146 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: audit 2026-03-06T21:16:46.392677+0000 mon.vm03 (mon.0) 147 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: audit 2026-03-06T21:16:46.392677+0000 mon.vm03 (mon.0) 147 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: audit 2026-03-06T21:16:46.397221+0000 mon.vm03 (mon.0) 148 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: audit 2026-03-06T21:16:46.397221+0000 mon.vm03 (mon.0) 148 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: audit 2026-03-06T21:16:46.400148+0000 mon.vm03 (mon.0) 149 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: audit 2026-03-06T21:16:46.400148+0000 mon.vm03 (mon.0) 149 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: cephadm 2026-03-06T21:16:46.580917+0000 mgr.vm03.uwuzgl (mgr.14168) 26 : cephadm [INF] Deploying daemon prometheus.vm03 on vm03 2026-03-06T22:16:47.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:47 vm03 bash[17055]: cephadm 2026-03-06T21:16:46.580917+0000 mgr.vm03.uwuzgl (mgr.14168) 26 : cephadm [INF] Deploying daemon prometheus.vm03 on vm03 2026-03-06T22:16:48.982 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:16:49.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:48 vm03 bash[17055]: cluster 2026-03-06T21:16:47.865755+0000 mgr.vm03.uwuzgl (mgr.14168) 27 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:49.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:48 vm03 bash[17055]: cluster 2026-03-06T21:16:47.865755+0000 mgr.vm03.uwuzgl (mgr.14168) 27 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:49.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:48 vm03 bash[17055]: audit 2026-03-06T21:16:47.898900+0000 mon.vm03 (mon.0) 150 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:49.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:48 vm03 bash[17055]: audit 2026-03-06T21:16:47.898900+0000 mon.vm03 (mon.0) 150 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:49.437 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:16:49.438 INFO:teuthology.orchestra.run.vm03.stdout:[{"addr": "192.168.123.103", "hostname": "vm03", "labels": [], "status": ""}, {"addr": "192.168.123.108", "hostname": "vm08", "labels": [], "status": ""}] 2026-03-06T22:16:49.540 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-06T22:16:49.540 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd crush tunables default 2026-03-06T22:16:50.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:49 vm03 bash[17055]: audit 2026-03-06T21:16:49.431433+0000 mgr.vm03.uwuzgl (mgr.14168) 28 : audit [DBG] from='client.14195 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:16:50.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:49 vm03 bash[17055]: audit 2026-03-06T21:16:49.431433+0000 mgr.vm03.uwuzgl (mgr.14168) 28 : audit [DBG] from='client.14195 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:16:51.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:50 vm03 bash[17055]: cluster 2026-03-06T21:16:49.865971+0000 mgr.vm03.uwuzgl (mgr.14168) 29 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:51.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:50 vm03 bash[17055]: cluster 2026-03-06T21:16:49.865971+0000 mgr.vm03.uwuzgl (mgr.14168) 29 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:53.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:52 vm03 bash[17055]: cluster 2026-03-06T21:16:51.866134+0000 mgr.vm03.uwuzgl (mgr.14168) 30 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:53.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:52 vm03 bash[17055]: cluster 2026-03-06T21:16:51.866134+0000 mgr.vm03.uwuzgl (mgr.14168) 30 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:54.375 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:16:55.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:54 vm03 bash[17055]: cluster 2026-03-06T21:16:53.866350+0000 mgr.vm03.uwuzgl (mgr.14168) 31 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:55.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:54 vm03 bash[17055]: cluster 2026-03-06T21:16:53.866350+0000 mgr.vm03.uwuzgl (mgr.14168) 31 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:16:55.469 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:55 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:16:55.469 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:55 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:16:55.929 INFO:teuthology.orchestra.run.vm03.stderr:adjusted tunables profile to default 2026-03-06T22:16:56.006 INFO:tasks.cephadm:Adding mon.vm03 on vm03 2026-03-06T22:16:56.006 INFO:tasks.cephadm:Adding mon.vm08 on vm08 2026-03-06T22:16:56.006 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph orch apply mon '2;vm03:192.168.123.103=vm03;vm08:192.168.123.108=vm08' 2026-03-06T22:16:56.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:55 vm03 bash[17055]: audit 2026-03-06T21:16:55.230504+0000 mon.vm03 (mon.0) 151 : audit [INF] from='client.? 192.168.123.103:0/1528358658' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-06T22:16:56.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:55 vm03 bash[17055]: audit 2026-03-06T21:16:55.230504+0000 mon.vm03 (mon.0) 151 : audit [INF] from='client.? 192.168.123.103:0/1528358658' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-06T22:16:56.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:55 vm03 bash[17055]: audit 2026-03-06T21:16:55.490634+0000 mon.vm03 (mon.0) 152 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:56.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:55 vm03 bash[17055]: audit 2026-03-06T21:16:55.490634+0000 mon.vm03 (mon.0) 152 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:56.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:55 vm03 bash[17055]: audit 2026-03-06T21:16:55.493565+0000 mon.vm03 (mon.0) 153 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:56.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:55 vm03 bash[17055]: audit 2026-03-06T21:16:55.493565+0000 mon.vm03 (mon.0) 153 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:56.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:55 vm03 bash[17055]: audit 2026-03-06T21:16:55.496123+0000 mon.vm03 (mon.0) 154 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:56.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:55 vm03 bash[17055]: audit 2026-03-06T21:16:55.496123+0000 mon.vm03 (mon.0) 154 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' 2026-03-06T22:16:56.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:55 vm03 bash[17055]: audit 2026-03-06T21:16:55.497455+0000 mon.vm03 (mon.0) 155 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr module enable", "module": "prometheus"}]: dispatch 2026-03-06T22:16:56.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:55 vm03 bash[17055]: audit 2026-03-06T21:16:55.497455+0000 mon.vm03 (mon.0) 155 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr module enable", "module": "prometheus"}]: dispatch 2026-03-06T22:16:57.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:56 vm03 bash[17055]: audit 2026-03-06T21:16:55.923575+0000 mon.vm03 (mon.0) 156 : audit [INF] from='client.? 192.168.123.103:0/1528358658' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-06T22:16:57.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:56 vm03 bash[17055]: audit 2026-03-06T21:16:55.923575+0000 mon.vm03 (mon.0) 156 : audit [INF] from='client.? 192.168.123.103:0/1528358658' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-06T22:16:57.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:56 vm03 bash[17055]: cluster 2026-03-06T21:16:55.926030+0000 mon.vm03 (mon.0) 157 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-06T22:16:57.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:56 vm03 bash[17055]: cluster 2026-03-06T21:16:55.926030+0000 mon.vm03 (mon.0) 157 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-06T22:16:57.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:56 vm03 bash[17055]: audit 2026-03-06T21:16:56.500186+0000 mon.vm03 (mon.0) 158 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "mgr module enable", "module": "prometheus"}]': finished 2026-03-06T22:16:57.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:56 vm03 bash[17055]: audit 2026-03-06T21:16:56.500186+0000 mon.vm03 (mon.0) 158 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "mgr module enable", "module": "prometheus"}]': finished 2026-03-06T22:16:57.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:56 vm03 bash[17055]: cluster 2026-03-06T21:16:56.502924+0000 mon.vm03 (mon.0) 159 : cluster [DBG] mgrmap e14: vm03.uwuzgl(active, since 38s) 2026-03-06T22:16:57.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:16:56 vm03 bash[17055]: cluster 2026-03-06T21:16:56.502924+0000 mon.vm03 (mon.0) 159 : cluster [DBG] mgrmap e14: vm03.uwuzgl(active, since 38s) 2026-03-06T22:16:57.259 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T22:16:58.280 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: cluster 2026-03-06T21:17:06.010808+0000 mon.vm03 (mon.0) 160 : cluster [INF] Active manager daemon vm03.uwuzgl restarted 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: cluster 2026-03-06T21:17:06.010808+0000 mon.vm03 (mon.0) 160 : cluster [INF] Active manager daemon vm03.uwuzgl restarted 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: cluster 2026-03-06T21:17:06.011195+0000 mon.vm03 (mon.0) 161 : cluster [INF] Activating manager daemon vm03.uwuzgl 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: cluster 2026-03-06T21:17:06.011195+0000 mon.vm03 (mon.0) 161 : cluster [INF] Activating manager daemon vm03.uwuzgl 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: cluster 2026-03-06T21:17:06.015824+0000 mon.vm03 (mon.0) 162 : cluster [DBG] osdmap e5: 0 total, 0 up, 0 in 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: cluster 2026-03-06T21:17:06.015824+0000 mon.vm03 (mon.0) 162 : cluster [DBG] osdmap e5: 0 total, 0 up, 0 in 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: cluster 2026-03-06T21:17:06.015919+0000 mon.vm03 (mon.0) 163 : cluster [DBG] mgrmap e15: vm03.uwuzgl(active, starting, since 0.00482585s) 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: cluster 2026-03-06T21:17:06.015919+0000 mon.vm03 (mon.0) 163 : cluster [DBG] mgrmap e15: vm03.uwuzgl(active, starting, since 0.00482585s) 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: audit 2026-03-06T21:17:06.019342+0000 mon.vm03 (mon.0) 164 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: audit 2026-03-06T21:17:06.019342+0000 mon.vm03 (mon.0) 164 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: audit 2026-03-06T21:17:06.020564+0000 mon.vm03 (mon.0) 165 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr metadata", "who": "vm03.uwuzgl", "id": "vm03.uwuzgl"}]: dispatch 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: audit 2026-03-06T21:17:06.020564+0000 mon.vm03 (mon.0) 165 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr metadata", "who": "vm03.uwuzgl", "id": "vm03.uwuzgl"}]: dispatch 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: audit 2026-03-06T21:17:06.021686+0000 mon.vm03 (mon.0) 166 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: audit 2026-03-06T21:17:06.021686+0000 mon.vm03 (mon.0) 166 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: audit 2026-03-06T21:17:06.022015+0000 mon.vm03 (mon.0) 167 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: audit 2026-03-06T21:17:06.022015+0000 mon.vm03 (mon.0) 167 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: audit 2026-03-06T21:17:06.022319+0000 mon.vm03 (mon.0) 168 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: audit 2026-03-06T21:17:06.022319+0000 mon.vm03 (mon.0) 168 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: cluster 2026-03-06T21:17:06.026512+0000 mon.vm03 (mon.0) 169 : cluster [INF] Manager daemon vm03.uwuzgl is now available 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: cluster 2026-03-06T21:17:06.026512+0000 mon.vm03 (mon.0) 169 : cluster [INF] Manager daemon vm03.uwuzgl is now available 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: audit 2026-03-06T21:17:06.042326+0000 mon.vm03 (mon.0) 170 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: audit 2026-03-06T21:17:06.042326+0000 mon.vm03 (mon.0) 170 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:06.351 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: audit 2026-03-06T21:17:06.053849+0000 mon.vm03 (mon.0) 171 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:17:06.352 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: audit 2026-03-06T21:17:06.053849+0000 mon.vm03 (mon.0) 171 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:17:06.352 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: audit 2026-03-06T21:17:06.055847+0000 mon.vm03 (mon.0) 172 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:17:06.352 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:06 vm03 bash[17055]: audit 2026-03-06T21:17:06.055847+0000 mon.vm03 (mon.0) 172 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:17:07.031 INFO:teuthology.orchestra.run.vm08.stdout:Scheduled mon update... 2026-03-06T22:17:07.107 DEBUG:teuthology.orchestra.run.vm08:mon.vm08> sudo journalctl -f -n 0 -u ceph-894e000c-19a1-11f1-8dbe-23b24380a082@mon.vm08.service 2026-03-06T22:17:07.108 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T22:17:07.108 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph mon dump -f json 2026-03-06T22:17:07.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:07 vm03 bash[17055]: audit 2026-03-06T21:17:06.069266+0000 mon.vm03 (mon.0) 173 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/mirror_snapshot_schedule"}]: dispatch 2026-03-06T22:17:07.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:07 vm03 bash[17055]: audit 2026-03-06T21:17:06.069266+0000 mon.vm03 (mon.0) 173 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/mirror_snapshot_schedule"}]: dispatch 2026-03-06T22:17:07.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:07 vm03 bash[17055]: audit 2026-03-06T21:17:06.116778+0000 mon.vm03 (mon.0) 174 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/trash_purge_schedule"}]: dispatch 2026-03-06T22:17:07.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:07 vm03 bash[17055]: audit 2026-03-06T21:17:06.116778+0000 mon.vm03 (mon.0) 174 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/trash_purge_schedule"}]: dispatch 2026-03-06T22:17:07.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:07 vm03 bash[17055]: audit 2026-03-06T21:17:06.731528+0000 mon.vm03 (mon.0) 175 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:07.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:07 vm03 bash[17055]: audit 2026-03-06T21:17:06.731528+0000 mon.vm03 (mon.0) 175 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:07.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:07 vm03 bash[17055]: cluster 2026-03-06T21:17:07.019471+0000 mon.vm03 (mon.0) 176 : cluster [DBG] mgrmap e16: vm03.uwuzgl(active, since 1.00838s) 2026-03-06T22:17:07.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:07 vm03 bash[17055]: cluster 2026-03-06T21:17:07.019471+0000 mon.vm03 (mon.0) 176 : cluster [DBG] mgrmap e16: vm03.uwuzgl(active, since 1.00838s) 2026-03-06T22:17:07.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:07 vm03 bash[17055]: audit 2026-03-06T21:17:07.025499+0000 mon.vm03 (mon.0) 177 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:07.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:07 vm03 bash[17055]: audit 2026-03-06T21:17:07.025499+0000 mon.vm03 (mon.0) 177 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:08.399 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T22:17:08.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:08 vm03 bash[17055]: cephadm 2026-03-06T21:17:07.022635+0000 mgr.vm03.uwuzgl (mgr.14199) 2 : cephadm [INF] Saving service mon spec with placement vm03:192.168.123.103=vm03;vm08:192.168.123.108=vm08;count:2 2026-03-06T22:17:08.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:08 vm03 bash[17055]: cephadm 2026-03-06T21:17:07.022635+0000 mgr.vm03.uwuzgl (mgr.14199) 2 : cephadm [INF] Saving service mon spec with placement vm03:192.168.123.103=vm03;vm08:192.168.123.108=vm08;count:2 2026-03-06T22:17:08.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:08 vm03 bash[17055]: cephadm 2026-03-06T21:17:07.122897+0000 mgr.vm03.uwuzgl (mgr.14199) 3 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Bus STARTING 2026-03-06T22:17:08.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:08 vm03 bash[17055]: cephadm 2026-03-06T21:17:07.122897+0000 mgr.vm03.uwuzgl (mgr.14199) 3 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Bus STARTING 2026-03-06T22:17:08.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:08 vm03 bash[17055]: cephadm 2026-03-06T21:17:07.233059+0000 mgr.vm03.uwuzgl (mgr.14199) 4 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Serving on https://192.168.123.103:7150 2026-03-06T22:17:08.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:08 vm03 bash[17055]: cephadm 2026-03-06T21:17:07.233059+0000 mgr.vm03.uwuzgl (mgr.14199) 4 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Serving on https://192.168.123.103:7150 2026-03-06T22:17:08.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:08 vm03 bash[17055]: cephadm 2026-03-06T21:17:07.233528+0000 mgr.vm03.uwuzgl (mgr.14199) 5 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Client ('192.168.123.103', 40822) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T22:17:08.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:08 vm03 bash[17055]: cephadm 2026-03-06T21:17:07.233528+0000 mgr.vm03.uwuzgl (mgr.14199) 5 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Client ('192.168.123.103', 40822) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T22:17:08.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:08 vm03 bash[17055]: cephadm 2026-03-06T21:17:07.334217+0000 mgr.vm03.uwuzgl (mgr.14199) 6 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Serving on http://192.168.123.103:8765 2026-03-06T22:17:08.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:08 vm03 bash[17055]: cephadm 2026-03-06T21:17:07.334217+0000 mgr.vm03.uwuzgl (mgr.14199) 6 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Serving on http://192.168.123.103:8765 2026-03-06T22:17:08.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:08 vm03 bash[17055]: cephadm 2026-03-06T21:17:07.334251+0000 mgr.vm03.uwuzgl (mgr.14199) 7 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Bus STARTED 2026-03-06T22:17:08.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:08 vm03 bash[17055]: cephadm 2026-03-06T21:17:07.334251+0000 mgr.vm03.uwuzgl (mgr.14199) 7 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Bus STARTED 2026-03-06T22:17:09.425 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-06T22:17:09.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:09 vm03 bash[17055]: cluster 2026-03-06T21:17:08.074512+0000 mon.vm03 (mon.0) 178 : cluster [DBG] mgrmap e17: vm03.uwuzgl(active, since 2s) 2026-03-06T22:17:09.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:09 vm03 bash[17055]: cluster 2026-03-06T21:17:08.074512+0000 mon.vm03 (mon.0) 178 : cluster [DBG] mgrmap e17: vm03.uwuzgl(active, since 2s) 2026-03-06T22:17:09.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:09 vm03 bash[17055]: audit 2026-03-06T21:17:08.110661+0000 mon.vm03 (mon.0) 179 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:09.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:09 vm03 bash[17055]: audit 2026-03-06T21:17:08.110661+0000 mon.vm03 (mon.0) 179 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:09.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:09 vm03 bash[17055]: audit 2026-03-06T21:17:08.672164+0000 mon.vm03 (mon.0) 180 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:09.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:09 vm03 bash[17055]: audit 2026-03-06T21:17:08.672164+0000 mon.vm03 (mon.0) 180 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:09.783 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-06T22:17:09.783 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T22:17:09.783 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"894e000c-19a1-11f1-8dbe-23b24380a082","modified":"2026-03-06T21:15:27.758447Z","created":"2026-03-06T21:15:27.758447Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm03","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:3300","nonce":0},{"type":"v1","addr":"192.168.123.103:6789","nonce":0}]},"addr":"192.168.123.103:6789/0","public_addr":"192.168.123.103:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T22:17:10.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:10 vm03 bash[17055]: audit 2026-03-06T21:17:09.777502+0000 mon.vm03 (mon.0) 181 : audit [DBG] from='client.? 192.168.123.108:0/174923989' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T22:17:10.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:10 vm03 bash[17055]: audit 2026-03-06T21:17:09.777502+0000 mon.vm03 (mon.0) 181 : audit [DBG] from='client.? 192.168.123.108:0/174923989' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T22:17:10.837 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T22:17:10.837 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph mon dump -f json 2026-03-06T22:17:12.689 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.conf 2026-03-06T22:17:13.151 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T22:17:13.151 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"894e000c-19a1-11f1-8dbe-23b24380a082","modified":"2026-03-06T21:15:27.758447Z","created":"2026-03-06T21:15:27.758447Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm03","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:3300","nonce":0},{"type":"v1","addr":"192.168.123.103:6789","nonce":0}]},"addr":"192.168.123.103:6789/0","public_addr":"192.168.123.103:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T22:17:13.151 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:11.702608+0000 mon.vm03 (mon.0) 182 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:11.702608+0000 mon.vm03 (mon.0) 182 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:11.704753+0000 mon.vm03 (mon.0) 183 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:11.704753+0000 mon.vm03 (mon.0) 183 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:11.706972+0000 mon.vm03 (mon.0) 184 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:11.706972+0000 mon.vm03 (mon.0) 184 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:11.708829+0000 mon.vm03 (mon.0) 185 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:11.708829+0000 mon.vm03 (mon.0) 185 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:11.709224+0000 mon.vm03 (mon.0) 186 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:11.709224+0000 mon.vm03 (mon.0) 186 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:11.902321+0000 mon.vm03 (mon.0) 187 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:11.902321+0000 mon.vm03 (mon.0) 187 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:11.905229+0000 mon.vm03 (mon.0) 188 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:11.905229+0000 mon.vm03 (mon.0) 188 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.437918+0000 mon.vm03 (mon.0) 189 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.437918+0000 mon.vm03 (mon.0) 189 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.440763+0000 mon.vm03 (mon.0) 190 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.440763+0000 mon.vm03 (mon.0) 190 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.441623+0000 mon.vm03 (mon.0) 191 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.441623+0000 mon.vm03 (mon.0) 191 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.442268+0000 mon.vm03 (mon.0) 192 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.442268+0000 mon.vm03 (mon.0) 192 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.442652+0000 mon.vm03 (mon.0) 193 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.442652+0000 mon.vm03 (mon.0) 193 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.443388+0000 mgr.vm03.uwuzgl (mgr.14199) 8 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.443388+0000 mgr.vm03.uwuzgl (mgr.14199) 8 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.443538+0000 mgr.vm03.uwuzgl (mgr.14199) 9 : cephadm [INF] Updating vm08:/etc/ceph/ceph.conf 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.443538+0000 mgr.vm03.uwuzgl (mgr.14199) 9 : cephadm [INF] Updating vm08:/etc/ceph/ceph.conf 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.479182+0000 mgr.vm03.uwuzgl (mgr.14199) 10 : cephadm [INF] Updating vm03:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.conf 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.479182+0000 mgr.vm03.uwuzgl (mgr.14199) 10 : cephadm [INF] Updating vm03:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.conf 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.481020+0000 mgr.vm03.uwuzgl (mgr.14199) 11 : cephadm [INF] Updating vm08:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.conf 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.481020+0000 mgr.vm03.uwuzgl (mgr.14199) 11 : cephadm [INF] Updating vm08:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.conf 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.509230+0000 mgr.vm03.uwuzgl (mgr.14199) 12 : cephadm [INF] Updating vm03:/etc/ceph/ceph.client.admin.keyring 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.509230+0000 mgr.vm03.uwuzgl (mgr.14199) 12 : cephadm [INF] Updating vm03:/etc/ceph/ceph.client.admin.keyring 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.510976+0000 mgr.vm03.uwuzgl (mgr.14199) 13 : cephadm [INF] Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.510976+0000 mgr.vm03.uwuzgl (mgr.14199) 13 : cephadm [INF] Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.538571+0000 mgr.vm03.uwuzgl (mgr.14199) 14 : cephadm [INF] Updating vm03:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.client.admin.keyring 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.538571+0000 mgr.vm03.uwuzgl (mgr.14199) 14 : cephadm [INF] Updating vm03:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.client.admin.keyring 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.540338+0000 mgr.vm03.uwuzgl (mgr.14199) 15 : cephadm [INF] Updating vm08:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.client.admin.keyring 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.540338+0000 mgr.vm03.uwuzgl (mgr.14199) 15 : cephadm [INF] Updating vm08:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.client.admin.keyring 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.573294+0000 mon.vm03 (mon.0) 194 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.573294+0000 mon.vm03 (mon.0) 194 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.576724+0000 mon.vm03 (mon.0) 195 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.576724+0000 mon.vm03 (mon.0) 195 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.579396+0000 mon.vm03 (mon.0) 196 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.579396+0000 mon.vm03 (mon.0) 196 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.581604+0000 mon.vm03 (mon.0) 197 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.581604+0000 mon.vm03 (mon.0) 197 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.584017+0000 mon.vm03 (mon.0) 198 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.584017+0000 mon.vm03 (mon.0) 198 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.585246+0000 mon.vm03 (mon.0) 199 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm08", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.585246+0000 mon.vm03 (mon.0) 199 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm08", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.586651+0000 mon.vm03 (mon.0) 200 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm08", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-06T22:17:13.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.586651+0000 mon.vm03 (mon.0) 200 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm08", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-06T22:17:13.192 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.587955+0000 mon.vm03 (mon.0) 201 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:13.192 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:12 vm03 bash[17055]: audit 2026-03-06T21:17:12.587955+0000 mon.vm03 (mon.0) 201 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:14.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:13 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.588427+0000 mgr.vm03.uwuzgl (mgr.14199) 16 : cephadm [INF] Deploying daemon ceph-exporter.vm08 on vm08 2026-03-06T22:17:14.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:13 vm03 bash[17055]: cephadm 2026-03-06T21:17:12.588427+0000 mgr.vm03.uwuzgl (mgr.14199) 16 : cephadm [INF] Deploying daemon ceph-exporter.vm08 on vm08 2026-03-06T22:17:14.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:13 vm03 bash[17055]: audit 2026-03-06T21:17:13.145592+0000 mon.vm03 (mon.0) 202 : audit [DBG] from='client.? 192.168.123.108:0/689693568' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T22:17:14.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:13 vm03 bash[17055]: audit 2026-03-06T21:17:13.145592+0000 mon.vm03 (mon.0) 202 : audit [DBG] from='client.? 192.168.123.108:0/689693568' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T22:17:14.403 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T22:17:14.403 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph mon dump -f json 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:13.974642+0000 mon.vm03 (mon.0) 203 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:13.974642+0000 mon.vm03 (mon.0) 203 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:13.977143+0000 mon.vm03 (mon.0) 204 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:13.977143+0000 mon.vm03 (mon.0) 204 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:13.979247+0000 mon.vm03 (mon.0) 205 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:13.979247+0000 mon.vm03 (mon.0) 205 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:13.981442+0000 mon.vm03 (mon.0) 206 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:13.981442+0000 mon.vm03 (mon.0) 206 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:13.982202+0000 mon.vm03 (mon.0) 207 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:13.982202+0000 mon.vm03 (mon.0) 207 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:13.983291+0000 mon.vm03 (mon.0) 208 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:13.983291+0000 mon.vm03 (mon.0) 208 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:13.984341+0000 mon.vm03 (mon.0) 209 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:13.984341+0000 mon.vm03 (mon.0) 209 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: cephadm 2026-03-06T21:17:13.985149+0000 mgr.vm03.uwuzgl (mgr.14199) 17 : cephadm [INF] Deploying daemon crash.vm08 on vm08 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: cephadm 2026-03-06T21:17:13.985149+0000 mgr.vm03.uwuzgl (mgr.14199) 17 : cephadm [INF] Deploying daemon crash.vm08 on vm08 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:14.756919+0000 mon.vm03 (mon.0) 210 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:14.756919+0000 mon.vm03 (mon.0) 210 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:14.759519+0000 mon.vm03 (mon.0) 211 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:14.759519+0000 mon.vm03 (mon.0) 211 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:14.761823+0000 mon.vm03 (mon.0) 212 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:14.761823+0000 mon.vm03 (mon.0) 212 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:14.763721+0000 mon.vm03 (mon.0) 213 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:14 vm03 bash[17055]: audit 2026-03-06T21:17:14.763721+0000 mon.vm03 (mon.0) 213 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:16 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: cephadm 2026-03-06T21:17:14.764637+0000 mgr.vm03.uwuzgl (mgr.14199) 18 : cephadm [INF] Deploying daemon node-exporter.vm08 on vm08 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: cephadm 2026-03-06T21:17:14.764637+0000 mgr.vm03.uwuzgl (mgr.14199) 18 : cephadm [INF] Deploying daemon node-exporter.vm08 on vm08 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:15.405981+0000 mon.vm03 (mon.0) 214 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:15.405981+0000 mon.vm03 (mon.0) 214 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:15.408256+0000 mon.vm03 (mon.0) 215 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:15.408256+0000 mon.vm03 (mon.0) 215 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:15.410476+0000 mon.vm03 (mon.0) 216 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:15.410476+0000 mon.vm03 (mon.0) 216 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:15.412532+0000 mon.vm03 (mon.0) 217 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:15.412532+0000 mon.vm03 (mon.0) 217 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:15.413709+0000 mon.vm03 (mon.0) 218 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm08.tdhano", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:15.413709+0000 mon.vm03 (mon.0) 218 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm08.tdhano", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:15.414666+0000 mon.vm03 (mon.0) 219 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.vm08.tdhano", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:15.414666+0000 mon.vm03 (mon.0) 219 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.vm08.tdhano", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:15.415733+0000 mon.vm03 (mon.0) 220 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:15.415733+0000 mon.vm03 (mon.0) 220 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:15.416188+0000 mon.vm03 (mon.0) 221 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:15.416188+0000 mon.vm03 (mon.0) 221 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: cephadm 2026-03-06T21:17:15.416656+0000 mgr.vm03.uwuzgl (mgr.14199) 19 : cephadm [INF] Deploying daemon mgr.vm08.tdhano on vm08 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: cephadm 2026-03-06T21:17:15.416656+0000 mgr.vm03.uwuzgl (mgr.14199) 19 : cephadm [INF] Deploying daemon mgr.vm08.tdhano on vm08 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:16.044300+0000 mon.vm03 (mon.0) 222 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:16.044300+0000 mon.vm03 (mon.0) 222 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:16.180592+0000 mon.vm03 (mon.0) 223 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:16.180592+0000 mon.vm03 (mon.0) 223 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:16.183298+0000 mon.vm03 (mon.0) 224 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:16.183298+0000 mon.vm03 (mon.0) 224 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:16.186183+0000 mon.vm03 (mon.0) 225 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:16.186183+0000 mon.vm03 (mon.0) 225 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:16.188801+0000 mon.vm03 (mon.0) 226 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:16.188801+0000 mon.vm03 (mon.0) 226 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:16.190184+0000 mon.vm03 (mon.0) 227 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:16.190184+0000 mon.vm03 (mon.0) 227 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T22:17:16.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:16.190709+0000 mon.vm03 (mon.0) 228 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:16.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:16 vm03 bash[17055]: audit 2026-03-06T21:17:16.190709+0000 mon.vm03 (mon.0) 228 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:17.098 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:16 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:17:17.098 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:16 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:17:17.098 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:16 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:17:17.407 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:17:17.408 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 systemd[1]: Started Ceph mon.vm08 for 894e000c-19a1-11f1-8dbe-23b24380a082. 2026-03-06T22:17:17.683 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.403+0000 7fd0a37eed80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.403+0000 7fd0a37eed80 0 ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable), process ceph-mon, pid 7 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.403+0000 7fd0a37eed80 0 pidfile_write: ignore empty --pid-file 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.403+0000 7fd0a37eed80 0 load: jerasure load: lrc 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: RocksDB version: 7.9.2 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Git sha 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Compile date 2026-03-06 13:52:12 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: DB SUMMARY 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: DB Session ID: CZO06GVNPYAQTSC7R08S 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: CURRENT file: CURRENT 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: MANIFEST file: MANIFEST-000005 size: 59 Bytes 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-vm08/store.db dir, Total Num: 0, files: 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-vm08/store.db: 000004.log size: 511 ; 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.error_if_exists: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.create_if_missing: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.paranoid_checks: 1 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.env: 0x555c9fa1eca0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.info_log: 0x555cb4daf820 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.statistics: (nil) 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.use_fsync: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_log_file_size: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.allow_fallocate: 1 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.use_direct_reads: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.db_log_dir: 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.wal_dir: 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.write_buffer_manager: 0x555cb4db3900 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-06T22:17:17.684 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.unordered_write: 0 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.row_cache: None 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.wal_filter: None 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.two_write_queues: 0 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.wal_compression: 0 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.atomic_flush: 0 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.log_readahead_size: 0 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_background_jobs: 2 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_background_compactions: -1 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_subcompactions: 1 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-06T22:17:17.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_open_files: -1 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_background_flushes: -1 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Compression algorithms supported: 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: kZSTD supported: 0 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: kXpressCompression supported: 0 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: kBZip2Compression supported: 0 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: kLZ4Compression supported: 1 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: kZlibCompression supported: 1 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: kSnappyCompression supported: 1 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-vm08/store.db/MANIFEST-000005 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.merge_operator: 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compaction_filter: None 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x555cb4daf460) 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cache_index_and_filter_blocks: 1 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: pin_top_level_index_and_filter: 1 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: index_type: 0 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: data_block_index_type: 0 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: index_shortening: 1 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: data_block_hash_table_util_ratio: 0.750000 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: checksum: 4 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: no_block_cache: 0 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: block_cache: 0x555cb4dd5350 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: block_cache_name: BinnedLRUCache 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: block_cache_options: 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: capacity : 536870912 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: num_shard_bits : 4 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: strict_capacity_limit : 0 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: high_pri_pool_ratio: 0.000 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: block_cache_compressed: (nil) 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: persistent_cache: (nil) 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: block_size: 4096 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: block_size_deviation: 10 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: block_restart_interval: 16 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: index_block_restart_interval: 1 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: metadata_block_size: 4096 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: partition_filters: 0 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: use_delta_encoding: 1 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: filter_policy: bloomfilter 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: whole_key_filtering: 1 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: verify_compression: 0 2026-03-06T22:17:17.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: read_amp_bytes_per_bit: 0 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: format_version: 5 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: enable_index_compression: 1 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: block_align: 0 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: max_auto_readahead_size: 262144 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: prepopulate_block_cache: 0 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: initial_auto_readahead_size: 8192 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: num_file_reads_for_auto_readahead: 2 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compression: NoCompression 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.num_levels: 7 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-06T22:17:17.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.inplace_update_support: 0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.bloom_locality: 0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.max_successive_merges: 0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.ttl: 2592000 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.enable_blob_files: false 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.min_blob_size: 0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-vm08/store.db/MANIFEST-000005 succeeded,manifest_file_number is 5, next_file_number is 7, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 3fd71cce-c903-4902-8ba3-629c8aec17e4 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772831837409298, "job": 1, "event": "recovery_started", "wal_files": [4]} 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.407+0000 7fd0a37eed80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #4 mode 2 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.423+0000 7fd0a37eed80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772831837424217, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 8, "file_size": 1643, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 1, "largest_seqno": 5, "table_properties": {"data_size": 523, "index_size": 31, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 115, "raw_average_key_size": 23, "raw_value_size": 401, "raw_average_value_size": 80, "num_data_blocks": 1, "num_entries": 5, "num_filter_entries": 5, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1772831837, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "3fd71cce-c903-4902-8ba3-629c8aec17e4", "db_session_id": "CZO06GVNPYAQTSC7R08S", "orig_file_number": 8, "seqno_to_time_mapping": "N/A"}} 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.423+0000 7fd0a37eed80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772831837424306, "job": 1, "event": "recovery_finished"} 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.423+0000 7fd0a37eed80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 10 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.423+0000 7fd0a37eed80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-vm08/store.db/000004.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.423+0000 7fd0a37eed80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x555cb4dd6e00 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.423+0000 7fd0a37eed80 4 rocksdb: DB pointer 0x555cb4dea000 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.423+0000 7fd0a37eed80 0 mon.vm08 does not exist in monmap, will attempt to join an existing cluster 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.423+0000 7fd0a37eed80 0 using public_addr v2:192.168.123.108:0/0 -> [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.427+0000 7fd0995b8640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.427+0000 7fd0995b8640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: ** DB Stats ** 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: ** Compaction Stats [default] ** 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: L0 1/0 1.60 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.1 0.01 0.00 1 0.015 0 0 0.0 0.0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Sum 1/0 1.60 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.1 0.01 0.00 1 0.015 0 0 0.0 0.0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.1 0.01 0.00 1 0.015 0 0 0.0 0.0 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: ** Compaction Stats [default] ** 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-06T22:17:17.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.1 0.01 0.00 1 0.015 0 0 0.0 0.0 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: AddFile(Total Files): cumulative 0, interval 0 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: AddFile(Keys): cumulative 0, interval 0 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Cumulative compaction: 0.00 GB write, 0.08 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Interval compaction: 0.00 GB write, 0.08 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Block cache BinnedLRUCache@0x555cb4dd5350#7 capacity: 512.00 MB usage: 0.86 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 6e-06 secs_since: 0 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: Block cache entry stats(count,size,portion): DataBlock(1,0.64 KB,0.00012219%) FilterBlock(1,0.11 KB,2.08616e-05%) IndexBlock(1,0.11 KB,2.08616e-05%) Misc(1,0.00 KB,0%) 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: ** File Read Latency Histogram By Level [default] ** 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.435+0000 7fd0a37eed80 0 starting mon.vm08 rank -1 at public addrs [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] at bind addrs [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon_data /var/lib/ceph/mon/ceph-vm08 fsid 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.435+0000 7fd0a37eed80 1 mon.vm08@-1(???) e0 preinit fsid 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.507+0000 7fd09c5be640 0 mon.vm08@-1(synchronizing).mds e1 new map 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.507+0000 7fd09c5be640 0 mon.vm08@-1(synchronizing).mds e1 print_map 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: e1 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: btime 2026-03-06T21:15:29:313048+0000 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: legacy client fscid: -1 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: No filesystems configured 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.507+0000 7fd09c5be640 1 mon.vm08@-1(synchronizing).osd e0 _set_cache_ratios kv ratio 0.25 inc ratio 0.375 full ratio 0.375 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.507+0000 7fd09c5be640 1 mon.vm08@-1(synchronizing).osd e0 register_cache_with_pcm pcm target: 2147483648 pcm max: 1020054732 pcm min: 134217728 inc_osd_cache size: 1 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.507+0000 7fd09c5be640 1 mon.vm08@-1(synchronizing).osd e1 e1: 0 total, 0 up, 0 in 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.507+0000 7fd09c5be640 1 mon.vm08@-1(synchronizing).osd e2 e2: 0 total, 0 up, 0 in 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.507+0000 7fd09c5be640 1 mon.vm08@-1(synchronizing).osd e3 e3: 0 total, 0 up, 0 in 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.507+0000 7fd09c5be640 1 mon.vm08@-1(synchronizing).osd e4 e4: 0 total, 0 up, 0 in 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.507+0000 7fd09c5be640 1 mon.vm08@-1(synchronizing).osd e5 e5: 0 total, 0 up, 0 in 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.507+0000 7fd09c5be640 0 mon.vm08@-1(synchronizing).osd e5 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.507+0000 7fd09c5be640 0 mon.vm08@-1(synchronizing).osd e5 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.507+0000 7fd09c5be640 0 mon.vm08@-1(synchronizing).osd e5 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.507+0000 7fd09c5be640 0 mon.vm08@-1(synchronizing).osd e5 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:16:55.923575+0000 mon.vm03 (mon.0) 156 : audit [INF] from='client.? 192.168.123.103:0/1528358658' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:16:55.923575+0000 mon.vm03 (mon.0) 156 : audit [INF] from='client.? 192.168.123.103:0/1528358658' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:16:55.926030+0000 mon.vm03 (mon.0) 157 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:16:55.926030+0000 mon.vm03 (mon.0) 157 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:16:56.500186+0000 mon.vm03 (mon.0) 158 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "mgr module enable", "module": "prometheus"}]': finished 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:16:56.500186+0000 mon.vm03 (mon.0) 158 : audit [INF] from='mgr.14168 192.168.123.103:0/658191111' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "mgr module enable", "module": "prometheus"}]': finished 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:16:56.502924+0000 mon.vm03 (mon.0) 159 : cluster [DBG] mgrmap e14: vm03.uwuzgl(active, since 38s) 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:16:56.502924+0000 mon.vm03 (mon.0) 159 : cluster [DBG] mgrmap e14: vm03.uwuzgl(active, since 38s) 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:17:06.010808+0000 mon.vm03 (mon.0) 160 : cluster [INF] Active manager daemon vm03.uwuzgl restarted 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:17:06.010808+0000 mon.vm03 (mon.0) 160 : cluster [INF] Active manager daemon vm03.uwuzgl restarted 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:17:06.011195+0000 mon.vm03 (mon.0) 161 : cluster [INF] Activating manager daemon vm03.uwuzgl 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:17:06.011195+0000 mon.vm03 (mon.0) 161 : cluster [INF] Activating manager daemon vm03.uwuzgl 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:17:06.015824+0000 mon.vm03 (mon.0) 162 : cluster [DBG] osdmap e5: 0 total, 0 up, 0 in 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:17:06.015824+0000 mon.vm03 (mon.0) 162 : cluster [DBG] osdmap e5: 0 total, 0 up, 0 in 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:17:06.015919+0000 mon.vm03 (mon.0) 163 : cluster [DBG] mgrmap e15: vm03.uwuzgl(active, starting, since 0.00482585s) 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:17:06.015919+0000 mon.vm03 (mon.0) 163 : cluster [DBG] mgrmap e15: vm03.uwuzgl(active, starting, since 0.00482585s) 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.019342+0000 mon.vm03 (mon.0) 164 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.019342+0000 mon.vm03 (mon.0) 164 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.020564+0000 mon.vm03 (mon.0) 165 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr metadata", "who": "vm03.uwuzgl", "id": "vm03.uwuzgl"}]: dispatch 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.020564+0000 mon.vm03 (mon.0) 165 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr metadata", "who": "vm03.uwuzgl", "id": "vm03.uwuzgl"}]: dispatch 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.021686+0000 mon.vm03 (mon.0) 166 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.021686+0000 mon.vm03 (mon.0) 166 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.022015+0000 mon.vm03 (mon.0) 167 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.022015+0000 mon.vm03 (mon.0) 167 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.022319+0000 mon.vm03 (mon.0) 168 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.022319+0000 mon.vm03 (mon.0) 168 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:17:06.026512+0000 mon.vm03 (mon.0) 169 : cluster [INF] Manager daemon vm03.uwuzgl is now available 2026-03-06T22:17:17.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:17:06.026512+0000 mon.vm03 (mon.0) 169 : cluster [INF] Manager daemon vm03.uwuzgl is now available 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.042326+0000 mon.vm03 (mon.0) 170 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.042326+0000 mon.vm03 (mon.0) 170 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.053849+0000 mon.vm03 (mon.0) 171 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.053849+0000 mon.vm03 (mon.0) 171 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.055847+0000 mon.vm03 (mon.0) 172 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.055847+0000 mon.vm03 (mon.0) 172 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.069266+0000 mon.vm03 (mon.0) 173 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/mirror_snapshot_schedule"}]: dispatch 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.069266+0000 mon.vm03 (mon.0) 173 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/mirror_snapshot_schedule"}]: dispatch 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.116778+0000 mon.vm03 (mon.0) 174 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/trash_purge_schedule"}]: dispatch 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.116778+0000 mon.vm03 (mon.0) 174 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm03.uwuzgl/trash_purge_schedule"}]: dispatch 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.731528+0000 mon.vm03 (mon.0) 175 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:06.731528+0000 mon.vm03 (mon.0) 175 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:17:07.019471+0000 mon.vm03 (mon.0) 176 : cluster [DBG] mgrmap e16: vm03.uwuzgl(active, since 1.00838s) 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:17:07.019471+0000 mon.vm03 (mon.0) 176 : cluster [DBG] mgrmap e16: vm03.uwuzgl(active, since 1.00838s) 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:07.025499+0000 mon.vm03 (mon.0) 177 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:07.025499+0000 mon.vm03 (mon.0) 177 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:07.022635+0000 mgr.vm03.uwuzgl (mgr.14199) 2 : cephadm [INF] Saving service mon spec with placement vm03:192.168.123.103=vm03;vm08:192.168.123.108=vm08;count:2 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:07.022635+0000 mgr.vm03.uwuzgl (mgr.14199) 2 : cephadm [INF] Saving service mon spec with placement vm03:192.168.123.103=vm03;vm08:192.168.123.108=vm08;count:2 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:07.122897+0000 mgr.vm03.uwuzgl (mgr.14199) 3 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Bus STARTING 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:07.122897+0000 mgr.vm03.uwuzgl (mgr.14199) 3 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Bus STARTING 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:07.233059+0000 mgr.vm03.uwuzgl (mgr.14199) 4 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Serving on https://192.168.123.103:7150 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:07.233059+0000 mgr.vm03.uwuzgl (mgr.14199) 4 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Serving on https://192.168.123.103:7150 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:07.233528+0000 mgr.vm03.uwuzgl (mgr.14199) 5 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Client ('192.168.123.103', 40822) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:07.233528+0000 mgr.vm03.uwuzgl (mgr.14199) 5 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Client ('192.168.123.103', 40822) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:07.334217+0000 mgr.vm03.uwuzgl (mgr.14199) 6 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Serving on http://192.168.123.103:8765 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:07.334217+0000 mgr.vm03.uwuzgl (mgr.14199) 6 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Serving on http://192.168.123.103:8765 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:07.334251+0000 mgr.vm03.uwuzgl (mgr.14199) 7 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Bus STARTED 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:07.334251+0000 mgr.vm03.uwuzgl (mgr.14199) 7 : cephadm [INF] [06/Mar/2026:21:17:07] ENGINE Bus STARTED 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:17:08.074512+0000 mon.vm03 (mon.0) 178 : cluster [DBG] mgrmap e17: vm03.uwuzgl(active, since 2s) 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cluster 2026-03-06T21:17:08.074512+0000 mon.vm03 (mon.0) 178 : cluster [DBG] mgrmap e17: vm03.uwuzgl(active, since 2s) 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:08.110661+0000 mon.vm03 (mon.0) 179 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:08.110661+0000 mon.vm03 (mon.0) 179 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:08.672164+0000 mon.vm03 (mon.0) 180 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:08.672164+0000 mon.vm03 (mon.0) 180 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:09.777502+0000 mon.vm03 (mon.0) 181 : audit [DBG] from='client.? 192.168.123.108:0/174923989' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:09.777502+0000 mon.vm03 (mon.0) 181 : audit [DBG] from='client.? 192.168.123.108:0/174923989' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:11.702608+0000 mon.vm03 (mon.0) 182 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:11.702608+0000 mon.vm03 (mon.0) 182 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:11.704753+0000 mon.vm03 (mon.0) 183 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:11.704753+0000 mon.vm03 (mon.0) 183 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:11.706972+0000 mon.vm03 (mon.0) 184 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:11.706972+0000 mon.vm03 (mon.0) 184 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:11.708829+0000 mon.vm03 (mon.0) 185 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:11.708829+0000 mon.vm03 (mon.0) 185 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:11.709224+0000 mon.vm03 (mon.0) 186 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:11.709224+0000 mon.vm03 (mon.0) 186 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:11.902321+0000 mon.vm03 (mon.0) 187 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:11.902321+0000 mon.vm03 (mon.0) 187 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:11.905229+0000 mon.vm03 (mon.0) 188 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:11.905229+0000 mon.vm03 (mon.0) 188 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.437918+0000 mon.vm03 (mon.0) 189 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.437918+0000 mon.vm03 (mon.0) 189 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.440763+0000 mon.vm03 (mon.0) 190 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.440763+0000 mon.vm03 (mon.0) 190 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.441623+0000 mon.vm03 (mon.0) 191 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.441623+0000 mon.vm03 (mon.0) 191 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.442268+0000 mon.vm03 (mon.0) 192 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:17.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.442268+0000 mon.vm03 (mon.0) 192 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:17 vm03 bash[17055]: cephadm 2026-03-06T21:17:16.191287+0000 mgr.vm03.uwuzgl (mgr.14199) 20 : cephadm [INF] Deploying daemon mon.vm08 on vm08 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:17 vm03 bash[17055]: cephadm 2026-03-06T21:17:16.191287+0000 mgr.vm03.uwuzgl (mgr.14199) 20 : cephadm [INF] Deploying daemon mon.vm08 on vm08 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:17 vm03 bash[17055]: audit 2026-03-06T21:17:17.187133+0000 mon.vm03 (mon.0) 229 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:17 vm03 bash[17055]: audit 2026-03-06T21:17:17.187133+0000 mon.vm03 (mon.0) 229 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:17 vm03 bash[17055]: audit 2026-03-06T21:17:17.190308+0000 mon.vm03 (mon.0) 230 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:17 vm03 bash[17055]: audit 2026-03-06T21:17:17.190308+0000 mon.vm03 (mon.0) 230 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:17 vm03 bash[17055]: audit 2026-03-06T21:17:17.193103+0000 mon.vm03 (mon.0) 231 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:17 vm03 bash[17055]: audit 2026-03-06T21:17:17.193103+0000 mon.vm03 (mon.0) 231 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:17 vm03 bash[17055]: audit 2026-03-06T21:17:17.196048+0000 mon.vm03 (mon.0) 232 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:17 vm03 bash[17055]: audit 2026-03-06T21:17:17.196048+0000 mon.vm03 (mon.0) 232 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:17 vm03 bash[17055]: audit 2026-03-06T21:17:17.199973+0000 mon.vm03 (mon.0) 233 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:17 vm03 bash[17055]: audit 2026-03-06T21:17:17.199973+0000 mon.vm03 (mon.0) 233 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:17 vm03 bash[17055]: audit 2026-03-06T21:17:17.212221+0000 mon.vm03 (mon.0) 234 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:17 vm03 bash[17055]: audit 2026-03-06T21:17:17.212221+0000 mon.vm03 (mon.0) 234 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.442652+0000 mon.vm03 (mon.0) 193 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.442652+0000 mon.vm03 (mon.0) 193 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.443388+0000 mgr.vm03.uwuzgl (mgr.14199) 8 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.443388+0000 mgr.vm03.uwuzgl (mgr.14199) 8 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.443538+0000 mgr.vm03.uwuzgl (mgr.14199) 9 : cephadm [INF] Updating vm08:/etc/ceph/ceph.conf 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.443538+0000 mgr.vm03.uwuzgl (mgr.14199) 9 : cephadm [INF] Updating vm08:/etc/ceph/ceph.conf 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.479182+0000 mgr.vm03.uwuzgl (mgr.14199) 10 : cephadm [INF] Updating vm03:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.conf 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.479182+0000 mgr.vm03.uwuzgl (mgr.14199) 10 : cephadm [INF] Updating vm03:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.conf 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.481020+0000 mgr.vm03.uwuzgl (mgr.14199) 11 : cephadm [INF] Updating vm08:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.conf 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.481020+0000 mgr.vm03.uwuzgl (mgr.14199) 11 : cephadm [INF] Updating vm08:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.conf 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.509230+0000 mgr.vm03.uwuzgl (mgr.14199) 12 : cephadm [INF] Updating vm03:/etc/ceph/ceph.client.admin.keyring 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.509230+0000 mgr.vm03.uwuzgl (mgr.14199) 12 : cephadm [INF] Updating vm03:/etc/ceph/ceph.client.admin.keyring 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.510976+0000 mgr.vm03.uwuzgl (mgr.14199) 13 : cephadm [INF] Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.510976+0000 mgr.vm03.uwuzgl (mgr.14199) 13 : cephadm [INF] Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.538571+0000 mgr.vm03.uwuzgl (mgr.14199) 14 : cephadm [INF] Updating vm03:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.client.admin.keyring 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.538571+0000 mgr.vm03.uwuzgl (mgr.14199) 14 : cephadm [INF] Updating vm03:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.client.admin.keyring 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.540338+0000 mgr.vm03.uwuzgl (mgr.14199) 15 : cephadm [INF] Updating vm08:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.client.admin.keyring 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.540338+0000 mgr.vm03.uwuzgl (mgr.14199) 15 : cephadm [INF] Updating vm08:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.client.admin.keyring 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.573294+0000 mon.vm03 (mon.0) 194 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.573294+0000 mon.vm03 (mon.0) 194 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.576724+0000 mon.vm03 (mon.0) 195 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.576724+0000 mon.vm03 (mon.0) 195 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.579396+0000 mon.vm03 (mon.0) 196 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.579396+0000 mon.vm03 (mon.0) 196 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.581604+0000 mon.vm03 (mon.0) 197 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.691 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.581604+0000 mon.vm03 (mon.0) 197 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.584017+0000 mon.vm03 (mon.0) 198 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.584017+0000 mon.vm03 (mon.0) 198 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.585246+0000 mon.vm03 (mon.0) 199 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm08", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.585246+0000 mon.vm03 (mon.0) 199 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm08", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.586651+0000 mon.vm03 (mon.0) 200 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm08", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.586651+0000 mon.vm03 (mon.0) 200 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm08", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.587955+0000 mon.vm03 (mon.0) 201 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:12.587955+0000 mon.vm03 (mon.0) 201 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.588427+0000 mgr.vm03.uwuzgl (mgr.14199) 16 : cephadm [INF] Deploying daemon ceph-exporter.vm08 on vm08 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:12.588427+0000 mgr.vm03.uwuzgl (mgr.14199) 16 : cephadm [INF] Deploying daemon ceph-exporter.vm08 on vm08 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:13.145592+0000 mon.vm03 (mon.0) 202 : audit [DBG] from='client.? 192.168.123.108:0/689693568' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:13.145592+0000 mon.vm03 (mon.0) 202 : audit [DBG] from='client.? 192.168.123.108:0/689693568' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:13.974642+0000 mon.vm03 (mon.0) 203 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:13.974642+0000 mon.vm03 (mon.0) 203 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:13.977143+0000 mon.vm03 (mon.0) 204 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:13.977143+0000 mon.vm03 (mon.0) 204 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:13.979247+0000 mon.vm03 (mon.0) 205 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:13.979247+0000 mon.vm03 (mon.0) 205 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:13.981442+0000 mon.vm03 (mon.0) 206 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:13.981442+0000 mon.vm03 (mon.0) 206 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:13.982202+0000 mon.vm03 (mon.0) 207 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:13.982202+0000 mon.vm03 (mon.0) 207 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:13.983291+0000 mon.vm03 (mon.0) 208 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:13.983291+0000 mon.vm03 (mon.0) 208 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:13.984341+0000 mon.vm03 (mon.0) 209 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:13.984341+0000 mon.vm03 (mon.0) 209 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:13.985149+0000 mgr.vm03.uwuzgl (mgr.14199) 17 : cephadm [INF] Deploying daemon crash.vm08 on vm08 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:13.985149+0000 mgr.vm03.uwuzgl (mgr.14199) 17 : cephadm [INF] Deploying daemon crash.vm08 on vm08 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:14.756919+0000 mon.vm03 (mon.0) 210 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:14.756919+0000 mon.vm03 (mon.0) 210 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:14.759519+0000 mon.vm03 (mon.0) 211 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:14.759519+0000 mon.vm03 (mon.0) 211 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:14.761823+0000 mon.vm03 (mon.0) 212 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:14.761823+0000 mon.vm03 (mon.0) 212 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:14.763721+0000 mon.vm03 (mon.0) 213 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:14.763721+0000 mon.vm03 (mon.0) 213 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:14.764637+0000 mgr.vm03.uwuzgl (mgr.14199) 18 : cephadm [INF] Deploying daemon node-exporter.vm08 on vm08 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:14.764637+0000 mgr.vm03.uwuzgl (mgr.14199) 18 : cephadm [INF] Deploying daemon node-exporter.vm08 on vm08 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:15.405981+0000 mon.vm03 (mon.0) 214 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:15.405981+0000 mon.vm03 (mon.0) 214 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:15.408256+0000 mon.vm03 (mon.0) 215 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:15.408256+0000 mon.vm03 (mon.0) 215 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:15.410476+0000 mon.vm03 (mon.0) 216 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:15.410476+0000 mon.vm03 (mon.0) 216 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:15.412532+0000 mon.vm03 (mon.0) 217 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:15.412532+0000 mon.vm03 (mon.0) 217 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:15.413709+0000 mon.vm03 (mon.0) 218 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm08.tdhano", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:15.413709+0000 mon.vm03 (mon.0) 218 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm08.tdhano", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:15.414666+0000 mon.vm03 (mon.0) 219 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.vm08.tdhano", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:15.414666+0000 mon.vm03 (mon.0) 219 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.vm08.tdhano", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:15.415733+0000 mon.vm03 (mon.0) 220 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:15.415733+0000 mon.vm03 (mon.0) 220 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:15.416188+0000 mon.vm03 (mon.0) 221 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:15.416188+0000 mon.vm03 (mon.0) 221 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:15.416656+0000 mgr.vm03.uwuzgl (mgr.14199) 19 : cephadm [INF] Deploying daemon mgr.vm08.tdhano on vm08 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:15.416656+0000 mgr.vm03.uwuzgl (mgr.14199) 19 : cephadm [INF] Deploying daemon mgr.vm08.tdhano on vm08 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:16.044300+0000 mon.vm03 (mon.0) 222 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:16.044300+0000 mon.vm03 (mon.0) 222 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:16.180592+0000 mon.vm03 (mon.0) 223 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:16.180592+0000 mon.vm03 (mon.0) 223 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:16.183298+0000 mon.vm03 (mon.0) 224 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:16.183298+0000 mon.vm03 (mon.0) 224 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.692 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:16.186183+0000 mon.vm03 (mon.0) 225 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:16.186183+0000 mon.vm03 (mon.0) 225 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:16.188801+0000 mon.vm03 (mon.0) 226 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:16.188801+0000 mon.vm03 (mon.0) 226 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:16.190184+0000 mon.vm03 (mon.0) 227 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:16.190184+0000 mon.vm03 (mon.0) 227 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:16.190709+0000 mon.vm03 (mon.0) 228 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:16.190709+0000 mon.vm03 (mon.0) 228 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:16.191287+0000 mgr.vm03.uwuzgl (mgr.14199) 20 : cephadm [INF] Deploying daemon mon.vm08 on vm08 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: cephadm 2026-03-06T21:17:16.191287+0000 mgr.vm03.uwuzgl (mgr.14199) 20 : cephadm [INF] Deploying daemon mon.vm08 on vm08 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:17.187133+0000 mon.vm03 (mon.0) 229 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:17.187133+0000 mon.vm03 (mon.0) 229 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:17.190308+0000 mon.vm03 (mon.0) 230 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:17.190308+0000 mon.vm03 (mon.0) 230 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:17.193103+0000 mon.vm03 (mon.0) 231 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:17.193103+0000 mon.vm03 (mon.0) 231 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:17.196048+0000 mon.vm03 (mon.0) 232 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:17.196048+0000 mon.vm03 (mon.0) 232 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:17.199973+0000 mon.vm03 (mon.0) 233 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:17.199973+0000 mon.vm03 (mon.0) 233 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:17.212221+0000 mon.vm03 (mon.0) 234 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: audit 2026-03-06T21:17:17.212221+0000 mon.vm03 (mon.0) 234 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:17:17.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:17 vm08 bash[20828]: debug 2026-03-06T21:17:17.511+0000 7fd09c5be640 1 mon.vm08@-1(synchronizing).paxosservice(auth 1..8) refresh upgraded, format 0 -> 3 2026-03-06T22:17:22.221 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm08/config 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:17.521913+0000 mon.vm03 (mon.0) 236 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:17.521913+0000 mon.vm03 (mon.0) 236 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:17.521970+0000 mon.vm03 (mon.0) 237 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:17.521970+0000 mon.vm03 (mon.0) 237 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:17.522160+0000 mon.vm03 (mon.0) 238 : cluster [INF] mon.vm03 calling monitor election 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:17.522160+0000 mon.vm03 (mon.0) 238 : cluster [INF] mon.vm03 calling monitor election 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:18.517234+0000 mon.vm03 (mon.0) 239 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:18.517234+0000 mon.vm03 (mon.0) 239 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:19.517504+0000 mon.vm03 (mon.0) 240 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:19.517504+0000 mon.vm03 (mon.0) 240 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:19.524855+0000 mon.vm08 (mon.1) 1 : cluster [INF] mon.vm08 calling monitor election 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:19.524855+0000 mon.vm08 (mon.1) 1 : cluster [INF] mon.vm08 calling monitor election 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:20.517493+0000 mon.vm03 (mon.0) 241 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:20.517493+0000 mon.vm03 (mon.0) 241 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:21.056222+0000 mon.vm03 (mon.0) 242 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:21.056222+0000 mon.vm03 (mon.0) 242 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:21.517676+0000 mon.vm03 (mon.0) 243 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:21.517676+0000 mon.vm03 (mon.0) 243 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:22.519574+0000 mon.vm03 (mon.0) 244 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:22.519574+0000 mon.vm03 (mon.0) 244 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.529791+0000 mon.vm03 (mon.0) 245 : cluster [INF] mon.vm03 is new leader, mons vm03,vm08 in quorum (ranks 0,1) 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.529791+0000 mon.vm03 (mon.0) 245 : cluster [INF] mon.vm03 is new leader, mons vm03,vm08 in quorum (ranks 0,1) 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533367+0000 mon.vm03 (mon.0) 246 : cluster [DBG] monmap epoch 2 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533367+0000 mon.vm03 (mon.0) 246 : cluster [DBG] monmap epoch 2 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533387+0000 mon.vm03 (mon.0) 247 : cluster [DBG] fsid 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533387+0000 mon.vm03 (mon.0) 247 : cluster [DBG] fsid 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533397+0000 mon.vm03 (mon.0) 248 : cluster [DBG] last_changed 2026-03-06T21:17:17.519059+0000 2026-03-06T22:17:22.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533397+0000 mon.vm03 (mon.0) 248 : cluster [DBG] last_changed 2026-03-06T21:17:17.519059+0000 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533406+0000 mon.vm03 (mon.0) 249 : cluster [DBG] created 2026-03-06T21:15:27.758447+0000 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533406+0000 mon.vm03 (mon.0) 249 : cluster [DBG] created 2026-03-06T21:15:27.758447+0000 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533416+0000 mon.vm03 (mon.0) 250 : cluster [DBG] min_mon_release 19 (squid) 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533416+0000 mon.vm03 (mon.0) 250 : cluster [DBG] min_mon_release 19 (squid) 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533425+0000 mon.vm03 (mon.0) 251 : cluster [DBG] election_strategy: 1 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533425+0000 mon.vm03 (mon.0) 251 : cluster [DBG] election_strategy: 1 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533435+0000 mon.vm03 (mon.0) 252 : cluster [DBG] 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.vm03 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533435+0000 mon.vm03 (mon.0) 252 : cluster [DBG] 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.vm03 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533444+0000 mon.vm03 (mon.0) 253 : cluster [DBG] 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.vm08 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533444+0000 mon.vm03 (mon.0) 253 : cluster [DBG] 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.vm08 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533750+0000 mon.vm03 (mon.0) 254 : cluster [DBG] fsmap 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533750+0000 mon.vm03 (mon.0) 254 : cluster [DBG] fsmap 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533770+0000 mon.vm03 (mon.0) 255 : cluster [DBG] osdmap e5: 0 total, 0 up, 0 in 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533770+0000 mon.vm03 (mon.0) 255 : cluster [DBG] osdmap e5: 0 total, 0 up, 0 in 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533889+0000 mon.vm03 (mon.0) 256 : cluster [DBG] mgrmap e17: vm03.uwuzgl(active, since 16s) 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533889+0000 mon.vm03 (mon.0) 256 : cluster [DBG] mgrmap e17: vm03.uwuzgl(active, since 16s) 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533967+0000 mon.vm03 (mon.0) 257 : cluster [INF] overall HEALTH_OK 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: cluster 2026-03-06T21:17:22.533967+0000 mon.vm03 (mon.0) 257 : cluster [INF] overall HEALTH_OK 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:22.540653+0000 mon.vm03 (mon.0) 258 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:22.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:22 vm03 bash[17055]: audit 2026-03-06T21:17:22.540653+0000 mon.vm03 (mon.0) 258 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:22.879 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 2 2026-03-06T22:17:22.879 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-06T22:17:22.879 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":2,"fsid":"894e000c-19a1-11f1-8dbe-23b24380a082","modified":"2026-03-06T21:17:17.519059Z","created":"2026-03-06T21:15:27.758447Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm03","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:3300","nonce":0},{"type":"v1","addr":"192.168.123.103:6789","nonce":0}]},"addr":"192.168.123.103:6789/0","public_addr":"192.168.123.103:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:17.521913+0000 mon.vm03 (mon.0) 236 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:17.521913+0000 mon.vm03 (mon.0) 236 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:17.521970+0000 mon.vm03 (mon.0) 237 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:17.521970+0000 mon.vm03 (mon.0) 237 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:17.522160+0000 mon.vm03 (mon.0) 238 : cluster [INF] mon.vm03 calling monitor election 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:17.522160+0000 mon.vm03 (mon.0) 238 : cluster [INF] mon.vm03 calling monitor election 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:18.517234+0000 mon.vm03 (mon.0) 239 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:18.517234+0000 mon.vm03 (mon.0) 239 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:19.517504+0000 mon.vm03 (mon.0) 240 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:19.517504+0000 mon.vm03 (mon.0) 240 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:19.524855+0000 mon.vm08 (mon.1) 1 : cluster [INF] mon.vm08 calling monitor election 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:19.524855+0000 mon.vm08 (mon.1) 1 : cluster [INF] mon.vm08 calling monitor election 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:20.517493+0000 mon.vm03 (mon.0) 241 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:20.517493+0000 mon.vm03 (mon.0) 241 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:21.056222+0000 mon.vm03 (mon.0) 242 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:21.056222+0000 mon.vm03 (mon.0) 242 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:21.517676+0000 mon.vm03 (mon.0) 243 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:21.517676+0000 mon.vm03 (mon.0) 243 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:22.519574+0000 mon.vm03 (mon.0) 244 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:22.519574+0000 mon.vm03 (mon.0) 244 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.529791+0000 mon.vm03 (mon.0) 245 : cluster [INF] mon.vm03 is new leader, mons vm03,vm08 in quorum (ranks 0,1) 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.529791+0000 mon.vm03 (mon.0) 245 : cluster [INF] mon.vm03 is new leader, mons vm03,vm08 in quorum (ranks 0,1) 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533367+0000 mon.vm03 (mon.0) 246 : cluster [DBG] monmap epoch 2 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533367+0000 mon.vm03 (mon.0) 246 : cluster [DBG] monmap epoch 2 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533387+0000 mon.vm03 (mon.0) 247 : cluster [DBG] fsid 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533387+0000 mon.vm03 (mon.0) 247 : cluster [DBG] fsid 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533397+0000 mon.vm03 (mon.0) 248 : cluster [DBG] last_changed 2026-03-06T21:17:17.519059+0000 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533397+0000 mon.vm03 (mon.0) 248 : cluster [DBG] last_changed 2026-03-06T21:17:17.519059+0000 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533406+0000 mon.vm03 (mon.0) 249 : cluster [DBG] created 2026-03-06T21:15:27.758447+0000 2026-03-06T22:17:22.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533406+0000 mon.vm03 (mon.0) 249 : cluster [DBG] created 2026-03-06T21:15:27.758447+0000 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533416+0000 mon.vm03 (mon.0) 250 : cluster [DBG] min_mon_release 19 (squid) 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533416+0000 mon.vm03 (mon.0) 250 : cluster [DBG] min_mon_release 19 (squid) 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533425+0000 mon.vm03 (mon.0) 251 : cluster [DBG] election_strategy: 1 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533425+0000 mon.vm03 (mon.0) 251 : cluster [DBG] election_strategy: 1 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533435+0000 mon.vm03 (mon.0) 252 : cluster [DBG] 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.vm03 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533435+0000 mon.vm03 (mon.0) 252 : cluster [DBG] 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.vm03 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533444+0000 mon.vm03 (mon.0) 253 : cluster [DBG] 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.vm08 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533444+0000 mon.vm03 (mon.0) 253 : cluster [DBG] 1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.vm08 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533750+0000 mon.vm03 (mon.0) 254 : cluster [DBG] fsmap 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533750+0000 mon.vm03 (mon.0) 254 : cluster [DBG] fsmap 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533770+0000 mon.vm03 (mon.0) 255 : cluster [DBG] osdmap e5: 0 total, 0 up, 0 in 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533770+0000 mon.vm03 (mon.0) 255 : cluster [DBG] osdmap e5: 0 total, 0 up, 0 in 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533889+0000 mon.vm03 (mon.0) 256 : cluster [DBG] mgrmap e17: vm03.uwuzgl(active, since 16s) 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533889+0000 mon.vm03 (mon.0) 256 : cluster [DBG] mgrmap e17: vm03.uwuzgl(active, since 16s) 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533967+0000 mon.vm03 (mon.0) 257 : cluster [INF] overall HEALTH_OK 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: cluster 2026-03-06T21:17:22.533967+0000 mon.vm03 (mon.0) 257 : cluster [INF] overall HEALTH_OK 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:22.540653+0000 mon.vm03 (mon.0) 258 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:22.901 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:22 vm08 bash[20828]: audit 2026-03-06T21:17:22.540653+0000 mon.vm03 (mon.0) 258 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:22.957 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-06T22:17:22.958 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph config generate-minimal-conf 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.792419+0000 mon.vm03 (mon.0) 259 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.792419+0000 mon.vm03 (mon.0) 259 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.797005+0000 mon.vm03 (mon.0) 260 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.797005+0000 mon.vm03 (mon.0) 260 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.797742+0000 mon.vm03 (mon.0) 261 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.797742+0000 mon.vm03 (mon.0) 261 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.798268+0000 mon.vm03 (mon.0) 262 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.798268+0000 mon.vm03 (mon.0) 262 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: cephadm 2026-03-06T21:17:22.798844+0000 mgr.vm03.uwuzgl (mgr.14199) 21 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: cephadm 2026-03-06T21:17:22.798844+0000 mgr.vm03.uwuzgl (mgr.14199) 21 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: cephadm 2026-03-06T21:17:22.798942+0000 mgr.vm03.uwuzgl (mgr.14199) 22 : cephadm [INF] Updating vm08:/etc/ceph/ceph.conf 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: cephadm 2026-03-06T21:17:22.798942+0000 mgr.vm03.uwuzgl (mgr.14199) 22 : cephadm [INF] Updating vm08:/etc/ceph/ceph.conf 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: cephadm 2026-03-06T21:17:22.832599+0000 mgr.vm03.uwuzgl (mgr.14199) 23 : cephadm [INF] Updating vm03:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.conf 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: cephadm 2026-03-06T21:17:22.832599+0000 mgr.vm03.uwuzgl (mgr.14199) 23 : cephadm [INF] Updating vm03:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.conf 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: cephadm 2026-03-06T21:17:22.834510+0000 mgr.vm03.uwuzgl (mgr.14199) 24 : cephadm [INF] Updating vm08:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.conf 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: cephadm 2026-03-06T21:17:22.834510+0000 mgr.vm03.uwuzgl (mgr.14199) 24 : cephadm [INF] Updating vm08:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.conf 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.872314+0000 mon.vm03 (mon.0) 263 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.872314+0000 mon.vm03 (mon.0) 263 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.873368+0000 mon.vm03 (mon.0) 264 : audit [DBG] from='client.? 192.168.123.108:0/962335813' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.873368+0000 mon.vm03 (mon.0) 264 : audit [DBG] from='client.? 192.168.123.108:0/962335813' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.876298+0000 mon.vm03 (mon.0) 265 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.876298+0000 mon.vm03 (mon.0) 265 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.880522+0000 mon.vm03 (mon.0) 266 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.880522+0000 mon.vm03 (mon.0) 266 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.884165+0000 mon.vm03 (mon.0) 267 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.884165+0000 mon.vm03 (mon.0) 267 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.888043+0000 mon.vm03 (mon.0) 268 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:22.888043+0000 mon.vm03 (mon.0) 268 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: cephadm 2026-03-06T21:17:22.899331+0000 mgr.vm03.uwuzgl (mgr.14199) 25 : cephadm [INF] Reconfiguring grafana.vm03 (dependencies changed)... 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: cephadm 2026-03-06T21:17:22.899331+0000 mgr.vm03.uwuzgl (mgr.14199) 25 : cephadm [INF] Reconfiguring grafana.vm03 (dependencies changed)... 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: cephadm 2026-03-06T21:17:22.933347+0000 mgr.vm03.uwuzgl (mgr.14199) 26 : cephadm [INF] Reconfiguring daemon grafana.vm03 on vm03 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: cephadm 2026-03-06T21:17:22.933347+0000 mgr.vm03.uwuzgl (mgr.14199) 26 : cephadm [INF] Reconfiguring daemon grafana.vm03 on vm03 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:23.517829+0000 mon.vm03 (mon.0) 269 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:23.517829+0000 mon.vm03 (mon.0) 269 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:23.685117+0000 mon.vm03 (mon.0) 270 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:23.685117+0000 mon.vm03 (mon.0) 270 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:23.688894+0000 mon.vm03 (mon.0) 271 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:23.688894+0000 mon.vm03 (mon.0) 271 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:23.689930+0000 mon.vm03 (mon.0) 272 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm03", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:23.689930+0000 mon.vm03 (mon.0) 272 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm03", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:23.690450+0000 mon.vm03 (mon.0) 273 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:23.915 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:23 vm03 bash[17055]: audit 2026-03-06T21:17:23.690450+0000 mon.vm03 (mon.0) 273 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.792419+0000 mon.vm03 (mon.0) 259 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.792419+0000 mon.vm03 (mon.0) 259 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.797005+0000 mon.vm03 (mon.0) 260 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.797005+0000 mon.vm03 (mon.0) 260 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.797742+0000 mon.vm03 (mon.0) 261 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.797742+0000 mon.vm03 (mon.0) 261 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.798268+0000 mon.vm03 (mon.0) 262 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.798268+0000 mon.vm03 (mon.0) 262 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: cephadm 2026-03-06T21:17:22.798844+0000 mgr.vm03.uwuzgl (mgr.14199) 21 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: cephadm 2026-03-06T21:17:22.798844+0000 mgr.vm03.uwuzgl (mgr.14199) 21 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: cephadm 2026-03-06T21:17:22.798942+0000 mgr.vm03.uwuzgl (mgr.14199) 22 : cephadm [INF] Updating vm08:/etc/ceph/ceph.conf 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: cephadm 2026-03-06T21:17:22.798942+0000 mgr.vm03.uwuzgl (mgr.14199) 22 : cephadm [INF] Updating vm08:/etc/ceph/ceph.conf 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: cephadm 2026-03-06T21:17:22.832599+0000 mgr.vm03.uwuzgl (mgr.14199) 23 : cephadm [INF] Updating vm03:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.conf 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: cephadm 2026-03-06T21:17:22.832599+0000 mgr.vm03.uwuzgl (mgr.14199) 23 : cephadm [INF] Updating vm03:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.conf 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: cephadm 2026-03-06T21:17:22.834510+0000 mgr.vm03.uwuzgl (mgr.14199) 24 : cephadm [INF] Updating vm08:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.conf 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: cephadm 2026-03-06T21:17:22.834510+0000 mgr.vm03.uwuzgl (mgr.14199) 24 : cephadm [INF] Updating vm08:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/config/ceph.conf 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.872314+0000 mon.vm03 (mon.0) 263 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.872314+0000 mon.vm03 (mon.0) 263 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.873368+0000 mon.vm03 (mon.0) 264 : audit [DBG] from='client.? 192.168.123.108:0/962335813' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.873368+0000 mon.vm03 (mon.0) 264 : audit [DBG] from='client.? 192.168.123.108:0/962335813' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.876298+0000 mon.vm03 (mon.0) 265 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.876298+0000 mon.vm03 (mon.0) 265 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.880522+0000 mon.vm03 (mon.0) 266 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.880522+0000 mon.vm03 (mon.0) 266 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.884165+0000 mon.vm03 (mon.0) 267 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.122 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.884165+0000 mon.vm03 (mon.0) 267 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.123 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.888043+0000 mon.vm03 (mon.0) 268 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.123 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:22.888043+0000 mon.vm03 (mon.0) 268 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.123 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: cephadm 2026-03-06T21:17:22.899331+0000 mgr.vm03.uwuzgl (mgr.14199) 25 : cephadm [INF] Reconfiguring grafana.vm03 (dependencies changed)... 2026-03-06T22:17:24.123 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: cephadm 2026-03-06T21:17:22.899331+0000 mgr.vm03.uwuzgl (mgr.14199) 25 : cephadm [INF] Reconfiguring grafana.vm03 (dependencies changed)... 2026-03-06T22:17:24.123 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: cephadm 2026-03-06T21:17:22.933347+0000 mgr.vm03.uwuzgl (mgr.14199) 26 : cephadm [INF] Reconfiguring daemon grafana.vm03 on vm03 2026-03-06T22:17:24.123 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: cephadm 2026-03-06T21:17:22.933347+0000 mgr.vm03.uwuzgl (mgr.14199) 26 : cephadm [INF] Reconfiguring daemon grafana.vm03 on vm03 2026-03-06T22:17:24.123 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:23.517829+0000 mon.vm03 (mon.0) 269 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:24.123 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:23.517829+0000 mon.vm03 (mon.0) 269 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:17:24.123 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:23.685117+0000 mon.vm03 (mon.0) 270 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.123 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:23.685117+0000 mon.vm03 (mon.0) 270 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.123 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:23.688894+0000 mon.vm03 (mon.0) 271 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.123 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:23.688894+0000 mon.vm03 (mon.0) 271 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:24.123 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:23.689930+0000 mon.vm03 (mon.0) 272 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm03", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T22:17:24.123 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:23.689930+0000 mon.vm03 (mon.0) 272 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm03", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T22:17:24.123 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:23.690450+0000 mon.vm03 (mon.0) 273 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:24.123 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:23 vm08 bash[20828]: audit 2026-03-06T21:17:23.690450+0000 mon.vm03 (mon.0) 273 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:25.380 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: cephadm 2026-03-06T21:17:23.689704+0000 mgr.vm03.uwuzgl (mgr.14199) 27 : cephadm [INF] Reconfiguring crash.vm03 (monmap changed)... 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: cephadm 2026-03-06T21:17:23.689704+0000 mgr.vm03.uwuzgl (mgr.14199) 27 : cephadm [INF] Reconfiguring crash.vm03 (monmap changed)... 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: cephadm 2026-03-06T21:17:23.690918+0000 mgr.vm03.uwuzgl (mgr.14199) 28 : cephadm [INF] Reconfiguring daemon crash.vm03 on vm03 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: cephadm 2026-03-06T21:17:23.690918+0000 mgr.vm03.uwuzgl (mgr.14199) 28 : cephadm [INF] Reconfiguring daemon crash.vm03 on vm03 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.115223+0000 mon.vm03 (mon.0) 274 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.115223+0000 mon.vm03 (mon.0) 274 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.120782+0000 mon.vm03 (mon.0) 275 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.120782+0000 mon.vm03 (mon.0) 275 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: cephadm 2026-03-06T21:17:24.121826+0000 mgr.vm03.uwuzgl (mgr.14199) 29 : cephadm [INF] Reconfiguring ceph-exporter.vm03 (monmap changed)... 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: cephadm 2026-03-06T21:17:24.121826+0000 mgr.vm03.uwuzgl (mgr.14199) 29 : cephadm [INF] Reconfiguring ceph-exporter.vm03 (monmap changed)... 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.122195+0000 mon.vm03 (mon.0) 276 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm03", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.122195+0000 mon.vm03 (mon.0) 276 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm03", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.122844+0000 mon.vm03 (mon.0) 277 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.122844+0000 mon.vm03 (mon.0) 277 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: cephadm 2026-03-06T21:17:24.123462+0000 mgr.vm03.uwuzgl (mgr.14199) 30 : cephadm [INF] Reconfiguring daemon ceph-exporter.vm03 on vm03 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: cephadm 2026-03-06T21:17:24.123462+0000 mgr.vm03.uwuzgl (mgr.14199) 30 : cephadm [INF] Reconfiguring daemon ceph-exporter.vm03 on vm03 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.508313+0000 mon.vm03 (mon.0) 278 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.508313+0000 mon.vm03 (mon.0) 278 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.513505+0000 mon.vm03 (mon.0) 279 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.513505+0000 mon.vm03 (mon.0) 279 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: cephadm 2026-03-06T21:17:24.514153+0000 mgr.vm03.uwuzgl (mgr.14199) 31 : cephadm [INF] Reconfiguring mon.vm03 (unknown last config time)... 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: cephadm 2026-03-06T21:17:24.514153+0000 mgr.vm03.uwuzgl (mgr.14199) 31 : cephadm [INF] Reconfiguring mon.vm03 (unknown last config time)... 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.514845+0000 mon.vm03 (mon.0) 280 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.514845+0000 mon.vm03 (mon.0) 280 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.515244+0000 mon.vm03 (mon.0) 281 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.515244+0000 mon.vm03 (mon.0) 281 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.515570+0000 mon.vm03 (mon.0) 282 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.515570+0000 mon.vm03 (mon.0) 282 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: cephadm 2026-03-06T21:17:24.516009+0000 mgr.vm03.uwuzgl (mgr.14199) 32 : cephadm [INF] Reconfiguring daemon mon.vm03 on vm03 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: cephadm 2026-03-06T21:17:24.516009+0000 mgr.vm03.uwuzgl (mgr.14199) 32 : cephadm [INF] Reconfiguring daemon mon.vm03 on vm03 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.877136+0000 mon.vm03 (mon.0) 283 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.877136+0000 mon.vm03 (mon.0) 283 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.881723+0000 mon.vm03 (mon.0) 284 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.381 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:25 vm03 bash[17055]: audit 2026-03-06T21:17:24.881723+0000 mon.vm03 (mon.0) 284 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: cephadm 2026-03-06T21:17:23.689704+0000 mgr.vm03.uwuzgl (mgr.14199) 27 : cephadm [INF] Reconfiguring crash.vm03 (monmap changed)... 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: cephadm 2026-03-06T21:17:23.689704+0000 mgr.vm03.uwuzgl (mgr.14199) 27 : cephadm [INF] Reconfiguring crash.vm03 (monmap changed)... 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: cephadm 2026-03-06T21:17:23.690918+0000 mgr.vm03.uwuzgl (mgr.14199) 28 : cephadm [INF] Reconfiguring daemon crash.vm03 on vm03 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: cephadm 2026-03-06T21:17:23.690918+0000 mgr.vm03.uwuzgl (mgr.14199) 28 : cephadm [INF] Reconfiguring daemon crash.vm03 on vm03 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.115223+0000 mon.vm03 (mon.0) 274 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.115223+0000 mon.vm03 (mon.0) 274 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.120782+0000 mon.vm03 (mon.0) 275 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.120782+0000 mon.vm03 (mon.0) 275 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: cephadm 2026-03-06T21:17:24.121826+0000 mgr.vm03.uwuzgl (mgr.14199) 29 : cephadm [INF] Reconfiguring ceph-exporter.vm03 (monmap changed)... 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: cephadm 2026-03-06T21:17:24.121826+0000 mgr.vm03.uwuzgl (mgr.14199) 29 : cephadm [INF] Reconfiguring ceph-exporter.vm03 (monmap changed)... 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.122195+0000 mon.vm03 (mon.0) 276 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm03", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.122195+0000 mon.vm03 (mon.0) 276 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm03", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.122844+0000 mon.vm03 (mon.0) 277 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.122844+0000 mon.vm03 (mon.0) 277 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: cephadm 2026-03-06T21:17:24.123462+0000 mgr.vm03.uwuzgl (mgr.14199) 30 : cephadm [INF] Reconfiguring daemon ceph-exporter.vm03 on vm03 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: cephadm 2026-03-06T21:17:24.123462+0000 mgr.vm03.uwuzgl (mgr.14199) 30 : cephadm [INF] Reconfiguring daemon ceph-exporter.vm03 on vm03 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.508313+0000 mon.vm03 (mon.0) 278 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.508313+0000 mon.vm03 (mon.0) 278 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.513505+0000 mon.vm03 (mon.0) 279 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.513505+0000 mon.vm03 (mon.0) 279 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: cephadm 2026-03-06T21:17:24.514153+0000 mgr.vm03.uwuzgl (mgr.14199) 31 : cephadm [INF] Reconfiguring mon.vm03 (unknown last config time)... 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: cephadm 2026-03-06T21:17:24.514153+0000 mgr.vm03.uwuzgl (mgr.14199) 31 : cephadm [INF] Reconfiguring mon.vm03 (unknown last config time)... 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.514845+0000 mon.vm03 (mon.0) 280 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.514845+0000 mon.vm03 (mon.0) 280 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.515244+0000 mon.vm03 (mon.0) 281 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.515244+0000 mon.vm03 (mon.0) 281 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.515570+0000 mon.vm03 (mon.0) 282 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.515570+0000 mon.vm03 (mon.0) 282 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: cephadm 2026-03-06T21:17:24.516009+0000 mgr.vm03.uwuzgl (mgr.14199) 32 : cephadm [INF] Reconfiguring daemon mon.vm03 on vm03 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: cephadm 2026-03-06T21:17:24.516009+0000 mgr.vm03.uwuzgl (mgr.14199) 32 : cephadm [INF] Reconfiguring daemon mon.vm03 on vm03 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.877136+0000 mon.vm03 (mon.0) 283 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.877136+0000 mon.vm03 (mon.0) 283 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.881723+0000 mon.vm03 (mon.0) 284 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:25 vm08 bash[20828]: audit 2026-03-06T21:17:24.881723+0000 mon.vm03 (mon.0) 284 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: cephadm 2026-03-06T21:17:24.882258+0000 mgr.vm03.uwuzgl (mgr.14199) 33 : cephadm [INF] Reconfiguring alertmanager.vm03 (dependencies changed)... 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: cephadm 2026-03-06T21:17:24.882258+0000 mgr.vm03.uwuzgl (mgr.14199) 33 : cephadm [INF] Reconfiguring alertmanager.vm03 (dependencies changed)... 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: cephadm 2026-03-06T21:17:24.885557+0000 mgr.vm03.uwuzgl (mgr.14199) 34 : cephadm [INF] Reconfiguring daemon alertmanager.vm03 on vm03 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: cephadm 2026-03-06T21:17:24.885557+0000 mgr.vm03.uwuzgl (mgr.14199) 34 : cephadm [INF] Reconfiguring daemon alertmanager.vm03 on vm03 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:25.447540+0000 mon.vm03 (mon.0) 285 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:25.447540+0000 mon.vm03 (mon.0) 285 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:25.454136+0000 mon.vm03 (mon.0) 286 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:25.454136+0000 mon.vm03 (mon.0) 286 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: cephadm 2026-03-06T21:17:25.454928+0000 mgr.vm03.uwuzgl (mgr.14199) 35 : cephadm [INF] Reconfiguring prometheus.vm03 (dependencies changed)... 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: cephadm 2026-03-06T21:17:25.454928+0000 mgr.vm03.uwuzgl (mgr.14199) 35 : cephadm [INF] Reconfiguring prometheus.vm03 (dependencies changed)... 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: cluster 2026-03-06T21:17:25.816219+0000 mon.vm03 (mon.0) 287 : cluster [DBG] Standby manager daemon vm08.tdhano started 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: cluster 2026-03-06T21:17:25.816219+0000 mon.vm03 (mon.0) 287 : cluster [DBG] Standby manager daemon vm08.tdhano started 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:25.817378+0000 mon.vm03 (mon.0) 288 : audit [DBG] from='mgr.? 192.168.123.108:0/1389973334' entity='mgr.vm08.tdhano' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm08.tdhano/crt"}]: dispatch 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:25.817378+0000 mon.vm03 (mon.0) 288 : audit [DBG] from='mgr.? 192.168.123.108:0/1389973334' entity='mgr.vm08.tdhano' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm08.tdhano/crt"}]: dispatch 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:25.817865+0000 mon.vm03 (mon.0) 289 : audit [DBG] from='mgr.? 192.168.123.108:0/1389973334' entity='mgr.vm08.tdhano' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:25.817865+0000 mon.vm03 (mon.0) 289 : audit [DBG] from='mgr.? 192.168.123.108:0/1389973334' entity='mgr.vm08.tdhano' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:25.818743+0000 mon.vm03 (mon.0) 290 : audit [DBG] from='mgr.? 192.168.123.108:0/1389973334' entity='mgr.vm08.tdhano' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm08.tdhano/key"}]: dispatch 2026-03-06T22:17:26.561 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:25.818743+0000 mon.vm03 (mon.0) 290 : audit [DBG] from='mgr.? 192.168.123.108:0/1389973334' entity='mgr.vm08.tdhano' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm08.tdhano/key"}]: dispatch 2026-03-06T22:17:26.562 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:25.819200+0000 mon.vm03 (mon.0) 291 : audit [DBG] from='mgr.? 192.168.123.108:0/1389973334' entity='mgr.vm08.tdhano' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-06T22:17:26.562 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:25.819200+0000 mon.vm03 (mon.0) 291 : audit [DBG] from='mgr.? 192.168.123.108:0/1389973334' entity='mgr.vm08.tdhano' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-06T22:17:26.562 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:26.224876+0000 mon.vm03 (mon.0) 292 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:26.562 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:26.224876+0000 mon.vm03 (mon.0) 292 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:26.562 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:26.228539+0000 mon.vm03 (mon.0) 293 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:26.562 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:26.228539+0000 mon.vm03 (mon.0) 293 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:26.562 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:26.229417+0000 mon.vm03 (mon.0) 294 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm03.uwuzgl", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T22:17:26.562 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:26.229417+0000 mon.vm03 (mon.0) 294 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm03.uwuzgl", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T22:17:26.562 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:26.229908+0000 mon.vm03 (mon.0) 295 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T22:17:26.562 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:26.229908+0000 mon.vm03 (mon.0) 295 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T22:17:26.562 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:26.230297+0000 mon.vm03 (mon.0) 296 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:26.562 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:26 vm03 bash[17055]: audit 2026-03-06T21:17:26.230297+0000 mon.vm03 (mon.0) 296 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:26.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: cephadm 2026-03-06T21:17:24.882258+0000 mgr.vm03.uwuzgl (mgr.14199) 33 : cephadm [INF] Reconfiguring alertmanager.vm03 (dependencies changed)... 2026-03-06T22:17:26.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: cephadm 2026-03-06T21:17:24.882258+0000 mgr.vm03.uwuzgl (mgr.14199) 33 : cephadm [INF] Reconfiguring alertmanager.vm03 (dependencies changed)... 2026-03-06T22:17:26.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: cephadm 2026-03-06T21:17:24.885557+0000 mgr.vm03.uwuzgl (mgr.14199) 34 : cephadm [INF] Reconfiguring daemon alertmanager.vm03 on vm03 2026-03-06T22:17:26.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: cephadm 2026-03-06T21:17:24.885557+0000 mgr.vm03.uwuzgl (mgr.14199) 34 : cephadm [INF] Reconfiguring daemon alertmanager.vm03 on vm03 2026-03-06T22:17:26.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:25.447540+0000 mon.vm03 (mon.0) 285 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:25.447540+0000 mon.vm03 (mon.0) 285 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:25.454136+0000 mon.vm03 (mon.0) 286 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:25.454136+0000 mon.vm03 (mon.0) 286 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: cephadm 2026-03-06T21:17:25.454928+0000 mgr.vm03.uwuzgl (mgr.14199) 35 : cephadm [INF] Reconfiguring prometheus.vm03 (dependencies changed)... 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: cephadm 2026-03-06T21:17:25.454928+0000 mgr.vm03.uwuzgl (mgr.14199) 35 : cephadm [INF] Reconfiguring prometheus.vm03 (dependencies changed)... 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: cluster 2026-03-06T21:17:25.816219+0000 mon.vm03 (mon.0) 287 : cluster [DBG] Standby manager daemon vm08.tdhano started 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: cluster 2026-03-06T21:17:25.816219+0000 mon.vm03 (mon.0) 287 : cluster [DBG] Standby manager daemon vm08.tdhano started 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:25.817378+0000 mon.vm03 (mon.0) 288 : audit [DBG] from='mgr.? 192.168.123.108:0/1389973334' entity='mgr.vm08.tdhano' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm08.tdhano/crt"}]: dispatch 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:25.817378+0000 mon.vm03 (mon.0) 288 : audit [DBG] from='mgr.? 192.168.123.108:0/1389973334' entity='mgr.vm08.tdhano' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm08.tdhano/crt"}]: dispatch 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:25.817865+0000 mon.vm03 (mon.0) 289 : audit [DBG] from='mgr.? 192.168.123.108:0/1389973334' entity='mgr.vm08.tdhano' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:25.817865+0000 mon.vm03 (mon.0) 289 : audit [DBG] from='mgr.? 192.168.123.108:0/1389973334' entity='mgr.vm08.tdhano' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:25.818743+0000 mon.vm03 (mon.0) 290 : audit [DBG] from='mgr.? 192.168.123.108:0/1389973334' entity='mgr.vm08.tdhano' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm08.tdhano/key"}]: dispatch 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:25.818743+0000 mon.vm03 (mon.0) 290 : audit [DBG] from='mgr.? 192.168.123.108:0/1389973334' entity='mgr.vm08.tdhano' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm08.tdhano/key"}]: dispatch 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:25.819200+0000 mon.vm03 (mon.0) 291 : audit [DBG] from='mgr.? 192.168.123.108:0/1389973334' entity='mgr.vm08.tdhano' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:25.819200+0000 mon.vm03 (mon.0) 291 : audit [DBG] from='mgr.? 192.168.123.108:0/1389973334' entity='mgr.vm08.tdhano' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:26.224876+0000 mon.vm03 (mon.0) 292 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:26.224876+0000 mon.vm03 (mon.0) 292 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:26.228539+0000 mon.vm03 (mon.0) 293 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:26.228539+0000 mon.vm03 (mon.0) 293 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:26.229417+0000 mon.vm03 (mon.0) 294 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm03.uwuzgl", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:26.229417+0000 mon.vm03 (mon.0) 294 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm03.uwuzgl", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:26.229908+0000 mon.vm03 (mon.0) 295 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:26.229908+0000 mon.vm03 (mon.0) 295 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:26.230297+0000 mon.vm03 (mon.0) 296 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:26.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:26 vm08 bash[20828]: audit 2026-03-06T21:17:26.230297+0000 mon.vm03 (mon.0) 296 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:27.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: cephadm 2026-03-06T21:17:25.646703+0000 mgr.vm03.uwuzgl (mgr.14199) 36 : cephadm [INF] Reconfiguring daemon prometheus.vm03 on vm03 2026-03-06T22:17:27.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: cephadm 2026-03-06T21:17:25.646703+0000 mgr.vm03.uwuzgl (mgr.14199) 36 : cephadm [INF] Reconfiguring daemon prometheus.vm03 on vm03 2026-03-06T22:17:27.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: cluster 2026-03-06T21:17:26.022697+0000 mgr.vm03.uwuzgl (mgr.14199) 37 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:27.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: cluster 2026-03-06T21:17:26.022697+0000 mgr.vm03.uwuzgl (mgr.14199) 37 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:27.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: cephadm 2026-03-06T21:17:26.229220+0000 mgr.vm03.uwuzgl (mgr.14199) 38 : cephadm [INF] Reconfiguring mgr.vm03.uwuzgl (unknown last config time)... 2026-03-06T22:17:27.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: cephadm 2026-03-06T21:17:26.229220+0000 mgr.vm03.uwuzgl (mgr.14199) 38 : cephadm [INF] Reconfiguring mgr.vm03.uwuzgl (unknown last config time)... 2026-03-06T22:17:27.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: cephadm 2026-03-06T21:17:26.230793+0000 mgr.vm03.uwuzgl (mgr.14199) 39 : cephadm [INF] Reconfiguring daemon mgr.vm03.uwuzgl on vm03 2026-03-06T22:17:27.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: cephadm 2026-03-06T21:17:26.230793+0000 mgr.vm03.uwuzgl (mgr.14199) 39 : cephadm [INF] Reconfiguring daemon mgr.vm03.uwuzgl on vm03 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: cluster 2026-03-06T21:17:26.464918+0000 mon.vm03 (mon.0) 297 : cluster [DBG] mgrmap e18: vm03.uwuzgl(active, since 20s), standbys: vm08.tdhano 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: cluster 2026-03-06T21:17:26.464918+0000 mon.vm03 (mon.0) 297 : cluster [DBG] mgrmap e18: vm03.uwuzgl(active, since 20s), standbys: vm08.tdhano 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:26.464981+0000 mon.vm03 (mon.0) 298 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr metadata", "who": "vm08.tdhano", "id": "vm08.tdhano"}]: dispatch 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:26.464981+0000 mon.vm03 (mon.0) 298 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr metadata", "who": "vm08.tdhano", "id": "vm08.tdhano"}]: dispatch 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:26.659815+0000 mon.vm03 (mon.0) 299 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:26.659815+0000 mon.vm03 (mon.0) 299 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:26.665441+0000 mon.vm03 (mon.0) 300 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:26.665441+0000 mon.vm03 (mon.0) 300 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:26.666301+0000 mon.vm03 (mon.0) 301 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm08.tdhano", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:26.666301+0000 mon.vm03 (mon.0) 301 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm08.tdhano", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:26.666758+0000 mon.vm03 (mon.0) 302 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:26.666758+0000 mon.vm03 (mon.0) 302 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:26.667098+0000 mon.vm03 (mon.0) 303 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:26.667098+0000 mon.vm03 (mon.0) 303 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:27.170983+0000 mon.vm03 (mon.0) 304 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:27.170983+0000 mon.vm03 (mon.0) 304 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:27.174535+0000 mon.vm03 (mon.0) 305 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:27.174535+0000 mon.vm03 (mon.0) 305 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:27.175163+0000 mon.vm03 (mon.0) 306 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:27.175163+0000 mon.vm03 (mon.0) 306 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:27.175569+0000 mon.vm03 (mon.0) 307 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:27.175569+0000 mon.vm03 (mon.0) 307 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:27.175924+0000 mon.vm03 (mon.0) 308 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:27.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:27 vm08 bash[20828]: audit 2026-03-06T21:17:27.175924+0000 mon.vm03 (mon.0) 308 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:27.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: cephadm 2026-03-06T21:17:25.646703+0000 mgr.vm03.uwuzgl (mgr.14199) 36 : cephadm [INF] Reconfiguring daemon prometheus.vm03 on vm03 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: cephadm 2026-03-06T21:17:25.646703+0000 mgr.vm03.uwuzgl (mgr.14199) 36 : cephadm [INF] Reconfiguring daemon prometheus.vm03 on vm03 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: cluster 2026-03-06T21:17:26.022697+0000 mgr.vm03.uwuzgl (mgr.14199) 37 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: cluster 2026-03-06T21:17:26.022697+0000 mgr.vm03.uwuzgl (mgr.14199) 37 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: cephadm 2026-03-06T21:17:26.229220+0000 mgr.vm03.uwuzgl (mgr.14199) 38 : cephadm [INF] Reconfiguring mgr.vm03.uwuzgl (unknown last config time)... 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: cephadm 2026-03-06T21:17:26.229220+0000 mgr.vm03.uwuzgl (mgr.14199) 38 : cephadm [INF] Reconfiguring mgr.vm03.uwuzgl (unknown last config time)... 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: cephadm 2026-03-06T21:17:26.230793+0000 mgr.vm03.uwuzgl (mgr.14199) 39 : cephadm [INF] Reconfiguring daemon mgr.vm03.uwuzgl on vm03 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: cephadm 2026-03-06T21:17:26.230793+0000 mgr.vm03.uwuzgl (mgr.14199) 39 : cephadm [INF] Reconfiguring daemon mgr.vm03.uwuzgl on vm03 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: cluster 2026-03-06T21:17:26.464918+0000 mon.vm03 (mon.0) 297 : cluster [DBG] mgrmap e18: vm03.uwuzgl(active, since 20s), standbys: vm08.tdhano 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: cluster 2026-03-06T21:17:26.464918+0000 mon.vm03 (mon.0) 297 : cluster [DBG] mgrmap e18: vm03.uwuzgl(active, since 20s), standbys: vm08.tdhano 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:26.464981+0000 mon.vm03 (mon.0) 298 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr metadata", "who": "vm08.tdhano", "id": "vm08.tdhano"}]: dispatch 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:26.464981+0000 mon.vm03 (mon.0) 298 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr metadata", "who": "vm08.tdhano", "id": "vm08.tdhano"}]: dispatch 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:26.659815+0000 mon.vm03 (mon.0) 299 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:26.659815+0000 mon.vm03 (mon.0) 299 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:26.665441+0000 mon.vm03 (mon.0) 300 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:26.665441+0000 mon.vm03 (mon.0) 300 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:26.666301+0000 mon.vm03 (mon.0) 301 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm08.tdhano", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:26.666301+0000 mon.vm03 (mon.0) 301 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm08.tdhano", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:26.666758+0000 mon.vm03 (mon.0) 302 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:26.666758+0000 mon.vm03 (mon.0) 302 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:26.667098+0000 mon.vm03 (mon.0) 303 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:26.667098+0000 mon.vm03 (mon.0) 303 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:27.170983+0000 mon.vm03 (mon.0) 304 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:27.170983+0000 mon.vm03 (mon.0) 304 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:27.174535+0000 mon.vm03 (mon.0) 305 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:27.174535+0000 mon.vm03 (mon.0) 305 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:27.175163+0000 mon.vm03 (mon.0) 306 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:27.175163+0000 mon.vm03 (mon.0) 306 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:27.175569+0000 mon.vm03 (mon.0) 307 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:27.175569+0000 mon.vm03 (mon.0) 307 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:27.175924+0000 mon.vm03 (mon.0) 308 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:27 vm03 bash[17055]: audit 2026-03-06T21:17:27.175924+0000 mon.vm03 (mon.0) 308 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:28.803 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:17:28.817 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: cephadm 2026-03-06T21:17:26.666111+0000 mgr.vm03.uwuzgl (mgr.14199) 40 : cephadm [INF] Reconfiguring mgr.vm08.tdhano (monmap changed)... 2026-03-06T22:17:28.817 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: cephadm 2026-03-06T21:17:26.666111+0000 mgr.vm03.uwuzgl (mgr.14199) 40 : cephadm [INF] Reconfiguring mgr.vm08.tdhano (monmap changed)... 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: cephadm 2026-03-06T21:17:26.667488+0000 mgr.vm03.uwuzgl (mgr.14199) 41 : cephadm [INF] Reconfiguring daemon mgr.vm08.tdhano on vm08 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: cephadm 2026-03-06T21:17:26.667488+0000 mgr.vm03.uwuzgl (mgr.14199) 41 : cephadm [INF] Reconfiguring daemon mgr.vm08.tdhano on vm08 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: cephadm 2026-03-06T21:17:27.175028+0000 mgr.vm03.uwuzgl (mgr.14199) 42 : cephadm [INF] Reconfiguring mon.vm08 (monmap changed)... 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: cephadm 2026-03-06T21:17:27.175028+0000 mgr.vm03.uwuzgl (mgr.14199) 42 : cephadm [INF] Reconfiguring mon.vm08 (monmap changed)... 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: cephadm 2026-03-06T21:17:27.176338+0000 mgr.vm03.uwuzgl (mgr.14199) 43 : cephadm [INF] Reconfiguring daemon mon.vm08 on vm08 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: cephadm 2026-03-06T21:17:27.176338+0000 mgr.vm03.uwuzgl (mgr.14199) 43 : cephadm [INF] Reconfiguring daemon mon.vm08 on vm08 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:27.535383+0000 mon.vm03 (mon.0) 309 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:27.535383+0000 mon.vm03 (mon.0) 309 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:27.538817+0000 mon.vm03 (mon.0) 310 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:27.538817+0000 mon.vm03 (mon.0) 310 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: cephadm 2026-03-06T21:17:27.539239+0000 mgr.vm03.uwuzgl (mgr.14199) 44 : cephadm [INF] Reconfiguring crash.vm08 (monmap changed)... 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: cephadm 2026-03-06T21:17:27.539239+0000 mgr.vm03.uwuzgl (mgr.14199) 44 : cephadm [INF] Reconfiguring crash.vm08 (monmap changed)... 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:27.539371+0000 mon.vm03 (mon.0) 311 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:27.539371+0000 mon.vm03 (mon.0) 311 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:27.539765+0000 mon.vm03 (mon.0) 312 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:27.539765+0000 mon.vm03 (mon.0) 312 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: cephadm 2026-03-06T21:17:27.540179+0000 mgr.vm03.uwuzgl (mgr.14199) 45 : cephadm [INF] Reconfiguring daemon crash.vm08 on vm08 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: cephadm 2026-03-06T21:17:27.540179+0000 mgr.vm03.uwuzgl (mgr.14199) 45 : cephadm [INF] Reconfiguring daemon crash.vm08 on vm08 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:27.877278+0000 mon.vm03 (mon.0) 313 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:27.877278+0000 mon.vm03 (mon.0) 313 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:27.880386+0000 mon.vm03 (mon.0) 314 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:27.880386+0000 mon.vm03 (mon.0) 314 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:27.881078+0000 mon.vm03 (mon.0) 315 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm08", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:27.881078+0000 mon.vm03 (mon.0) 315 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm08", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:27.881592+0000 mon.vm03 (mon.0) 316 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:27.881592+0000 mon.vm03 (mon.0) 316 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.242697+0000 mon.vm03 (mon.0) 317 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.242697+0000 mon.vm03 (mon.0) 317 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.247158+0000 mon.vm03 (mon.0) 318 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.247158+0000 mon.vm03 (mon.0) 318 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.249984+0000 mon.vm03 (mon.0) 319 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.249984+0000 mon.vm03 (mon.0) 319 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.251153+0000 mon.vm03 (mon.0) 320 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm03.local:3000"}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.251153+0000 mon.vm03 (mon.0) 320 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm03.local:3000"}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.254809+0000 mon.vm03 (mon.0) 321 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.254809+0000 mon.vm03 (mon.0) 321 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.263323+0000 mon.vm03 (mon.0) 322 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.263323+0000 mon.vm03 (mon.0) 322 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.264437+0000 mon.vm03 (mon.0) 323 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm03.local:9093"}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.264437+0000 mon.vm03 (mon.0) 323 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm03.local:9093"}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.268951+0000 mon.vm03 (mon.0) 324 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.268951+0000 mon.vm03 (mon.0) 324 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.275823+0000 mon.vm03 (mon.0) 325 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.275823+0000 mon.vm03 (mon.0) 325 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.276929+0000 mon.vm03 (mon.0) 326 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm03.local:9095"}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.276929+0000 mon.vm03 (mon.0) 326 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm03.local:9095"}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.280626+0000 mon.vm03 (mon.0) 327 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.280626+0000 mon.vm03 (mon.0) 327 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.312943+0000 mon.vm03 (mon.0) 328 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:17:28.818 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:28 vm03 bash[17055]: audit 2026-03-06T21:17:28.312943+0000 mon.vm03 (mon.0) 328 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: cephadm 2026-03-06T21:17:26.666111+0000 mgr.vm03.uwuzgl (mgr.14199) 40 : cephadm [INF] Reconfiguring mgr.vm08.tdhano (monmap changed)... 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: cephadm 2026-03-06T21:17:26.666111+0000 mgr.vm03.uwuzgl (mgr.14199) 40 : cephadm [INF] Reconfiguring mgr.vm08.tdhano (monmap changed)... 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: cephadm 2026-03-06T21:17:26.667488+0000 mgr.vm03.uwuzgl (mgr.14199) 41 : cephadm [INF] Reconfiguring daemon mgr.vm08.tdhano on vm08 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: cephadm 2026-03-06T21:17:26.667488+0000 mgr.vm03.uwuzgl (mgr.14199) 41 : cephadm [INF] Reconfiguring daemon mgr.vm08.tdhano on vm08 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: cephadm 2026-03-06T21:17:27.175028+0000 mgr.vm03.uwuzgl (mgr.14199) 42 : cephadm [INF] Reconfiguring mon.vm08 (monmap changed)... 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: cephadm 2026-03-06T21:17:27.175028+0000 mgr.vm03.uwuzgl (mgr.14199) 42 : cephadm [INF] Reconfiguring mon.vm08 (monmap changed)... 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: cephadm 2026-03-06T21:17:27.176338+0000 mgr.vm03.uwuzgl (mgr.14199) 43 : cephadm [INF] Reconfiguring daemon mon.vm08 on vm08 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: cephadm 2026-03-06T21:17:27.176338+0000 mgr.vm03.uwuzgl (mgr.14199) 43 : cephadm [INF] Reconfiguring daemon mon.vm08 on vm08 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:27.535383+0000 mon.vm03 (mon.0) 309 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:27.535383+0000 mon.vm03 (mon.0) 309 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:27.538817+0000 mon.vm03 (mon.0) 310 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:27.538817+0000 mon.vm03 (mon.0) 310 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: cephadm 2026-03-06T21:17:27.539239+0000 mgr.vm03.uwuzgl (mgr.14199) 44 : cephadm [INF] Reconfiguring crash.vm08 (monmap changed)... 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: cephadm 2026-03-06T21:17:27.539239+0000 mgr.vm03.uwuzgl (mgr.14199) 44 : cephadm [INF] Reconfiguring crash.vm08 (monmap changed)... 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:27.539371+0000 mon.vm03 (mon.0) 311 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:27.539371+0000 mon.vm03 (mon.0) 311 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:27.539765+0000 mon.vm03 (mon.0) 312 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:27.539765+0000 mon.vm03 (mon.0) 312 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: cephadm 2026-03-06T21:17:27.540179+0000 mgr.vm03.uwuzgl (mgr.14199) 45 : cephadm [INF] Reconfiguring daemon crash.vm08 on vm08 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: cephadm 2026-03-06T21:17:27.540179+0000 mgr.vm03.uwuzgl (mgr.14199) 45 : cephadm [INF] Reconfiguring daemon crash.vm08 on vm08 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:27.877278+0000 mon.vm03 (mon.0) 313 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:27.877278+0000 mon.vm03 (mon.0) 313 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:27.880386+0000 mon.vm03 (mon.0) 314 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:27.880386+0000 mon.vm03 (mon.0) 314 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:27.881078+0000 mon.vm03 (mon.0) 315 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm08", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:27.881078+0000 mon.vm03 (mon.0) 315 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm08", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:27.881592+0000 mon.vm03 (mon.0) 316 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:27.881592+0000 mon.vm03 (mon.0) 316 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.242697+0000 mon.vm03 (mon.0) 317 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.242697+0000 mon.vm03 (mon.0) 317 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.247158+0000 mon.vm03 (mon.0) 318 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.247158+0000 mon.vm03 (mon.0) 318 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.249984+0000 mon.vm03 (mon.0) 319 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.249984+0000 mon.vm03 (mon.0) 319 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.251153+0000 mon.vm03 (mon.0) 320 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm03.local:3000"}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.251153+0000 mon.vm03 (mon.0) 320 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm03.local:3000"}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.254809+0000 mon.vm03 (mon.0) 321 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.254809+0000 mon.vm03 (mon.0) 321 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.263323+0000 mon.vm03 (mon.0) 322 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.263323+0000 mon.vm03 (mon.0) 322 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.264437+0000 mon.vm03 (mon.0) 323 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm03.local:9093"}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.264437+0000 mon.vm03 (mon.0) 323 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm03.local:9093"}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.268951+0000 mon.vm03 (mon.0) 324 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.268951+0000 mon.vm03 (mon.0) 324 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.275823+0000 mon.vm03 (mon.0) 325 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T22:17:29.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.275823+0000 mon.vm03 (mon.0) 325 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T22:17:29.002 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.276929+0000 mon.vm03 (mon.0) 326 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm03.local:9095"}]: dispatch 2026-03-06T22:17:29.002 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.276929+0000 mon.vm03 (mon.0) 326 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm03.local:9095"}]: dispatch 2026-03-06T22:17:29.002 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.280626+0000 mon.vm03 (mon.0) 327 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.002 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.280626+0000 mon.vm03 (mon.0) 327 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:29.002 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.312943+0000 mon.vm03 (mon.0) 328 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:17:29.002 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:28 vm08 bash[20828]: audit 2026-03-06T21:17:28.312943+0000 mon.vm03 (mon.0) 328 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:17:29.170 INFO:teuthology.orchestra.run.vm03.stdout:# minimal ceph.conf for 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:17:29.170 INFO:teuthology.orchestra.run.vm03.stdout:[global] 2026-03-06T22:17:29.170 INFO:teuthology.orchestra.run.vm03.stdout: fsid = 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:17:29.170 INFO:teuthology.orchestra.run.vm03.stdout: mon_host = [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] 2026-03-06T22:17:29.232 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-06T22:17:29.232 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-06T22:17:29.232 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/etc/ceph/ceph.conf 2026-03-06T22:17:29.239 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-06T22:17:29.239 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T22:17:29.288 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T22:17:29.288 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/ceph.conf 2026-03-06T22:17:29.295 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T22:17:29.295 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T22:17:29.343 INFO:tasks.cephadm:Deploying OSDs... 2026-03-06T22:17:29.343 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-06T22:17:29.343 DEBUG:teuthology.orchestra.run.vm03:> dd if=/scratch_devs of=/dev/stdout 2026-03-06T22:17:29.345 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T22:17:29.345 DEBUG:teuthology.orchestra.run.vm03:> ls /dev/[sv]d? 2026-03-06T22:17:29.391 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vda 2026-03-06T22:17:29.392 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vdb 2026-03-06T22:17:29.392 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vdc 2026-03-06T22:17:29.392 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vdd 2026-03-06T22:17:29.392 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vde 2026-03-06T22:17:29.392 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-06T22:17:29.392 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-06T22:17:29.392 DEBUG:teuthology.orchestra.run.vm03:> stat /dev/vdb 2026-03-06T22:17:29.436 INFO:teuthology.orchestra.run.vm03.stdout: File: /dev/vdb 2026-03-06T22:17:29.436 INFO:teuthology.orchestra.run.vm03.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-06T22:17:29.436 INFO:teuthology.orchestra.run.vm03.stdout:Device: 5h/5d Inode: 24 Links: 1 Device type: fe,10 2026-03-06T22:17:29.436 INFO:teuthology.orchestra.run.vm03.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T22:17:29.436 INFO:teuthology.orchestra.run.vm03.stdout:Access: 2026-03-06 22:11:24.498585339 +0100 2026-03-06T22:17:29.436 INFO:teuthology.orchestra.run.vm03.stdout:Modify: 2026-03-06 22:11:22.718585339 +0100 2026-03-06T22:17:29.436 INFO:teuthology.orchestra.run.vm03.stdout:Change: 2026-03-06 22:11:22.718585339 +0100 2026-03-06T22:17:29.436 INFO:teuthology.orchestra.run.vm03.stdout: Birth: - 2026-03-06T22:17:29.436 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-06T22:17:29.482 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records in 2026-03-06T22:17:29.482 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records out 2026-03-06T22:17:29.482 INFO:teuthology.orchestra.run.vm03.stderr:512 bytes copied, 0.000104505 s, 4.9 MB/s 2026-03-06T22:17:29.483 DEBUG:teuthology.orchestra.run.vm03:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-06T22:17:29.529 DEBUG:teuthology.orchestra.run.vm03:> stat /dev/vdc 2026-03-06T22:17:29.572 INFO:teuthology.orchestra.run.vm03.stdout: File: /dev/vdc 2026-03-06T22:17:29.572 INFO:teuthology.orchestra.run.vm03.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-06T22:17:29.572 INFO:teuthology.orchestra.run.vm03.stdout:Device: 5h/5d Inode: 25 Links: 1 Device type: fe,20 2026-03-06T22:17:29.572 INFO:teuthology.orchestra.run.vm03.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T22:17:29.572 INFO:teuthology.orchestra.run.vm03.stdout:Access: 2026-03-06 22:11:24.510585339 +0100 2026-03-06T22:17:29.572 INFO:teuthology.orchestra.run.vm03.stdout:Modify: 2026-03-06 22:11:22.722585339 +0100 2026-03-06T22:17:29.572 INFO:teuthology.orchestra.run.vm03.stdout:Change: 2026-03-06 22:11:22.722585339 +0100 2026-03-06T22:17:29.572 INFO:teuthology.orchestra.run.vm03.stdout: Birth: - 2026-03-06T22:17:29.572 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-06T22:17:29.620 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records in 2026-03-06T22:17:29.620 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records out 2026-03-06T22:17:29.620 INFO:teuthology.orchestra.run.vm03.stderr:512 bytes copied, 0.000120386 s, 4.3 MB/s 2026-03-06T22:17:29.620 DEBUG:teuthology.orchestra.run.vm03:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-06T22:17:29.665 DEBUG:teuthology.orchestra.run.vm03:> stat /dev/vdd 2026-03-06T22:17:29.712 INFO:teuthology.orchestra.run.vm03.stdout: File: /dev/vdd 2026-03-06T22:17:29.712 INFO:teuthology.orchestra.run.vm03.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-06T22:17:29.712 INFO:teuthology.orchestra.run.vm03.stdout:Device: 5h/5d Inode: 26 Links: 1 Device type: fe,30 2026-03-06T22:17:29.712 INFO:teuthology.orchestra.run.vm03.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T22:17:29.712 INFO:teuthology.orchestra.run.vm03.stdout:Access: 2026-03-06 22:11:24.494585339 +0100 2026-03-06T22:17:29.712 INFO:teuthology.orchestra.run.vm03.stdout:Modify: 2026-03-06 22:11:22.762585339 +0100 2026-03-06T22:17:29.712 INFO:teuthology.orchestra.run.vm03.stdout:Change: 2026-03-06 22:11:22.762585339 +0100 2026-03-06T22:17:29.712 INFO:teuthology.orchestra.run.vm03.stdout: Birth: - 2026-03-06T22:17:29.712 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: cephadm 2026-03-06T21:17:27.880927+0000 mgr.vm03.uwuzgl (mgr.14199) 46 : cephadm [INF] Reconfiguring ceph-exporter.vm08 (monmap changed)... 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: cephadm 2026-03-06T21:17:27.880927+0000 mgr.vm03.uwuzgl (mgr.14199) 46 : cephadm [INF] Reconfiguring ceph-exporter.vm08 (monmap changed)... 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: cephadm 2026-03-06T21:17:27.881958+0000 mgr.vm03.uwuzgl (mgr.14199) 47 : cephadm [INF] Reconfiguring daemon ceph-exporter.vm08 on vm08 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: cephadm 2026-03-06T21:17:27.881958+0000 mgr.vm03.uwuzgl (mgr.14199) 47 : cephadm [INF] Reconfiguring daemon ceph-exporter.vm08 on vm08 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: cluster 2026-03-06T21:17:28.022843+0000 mgr.vm03.uwuzgl (mgr.14199) 48 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: cluster 2026-03-06T21:17:28.022843+0000 mgr.vm03.uwuzgl (mgr.14199) 48 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: audit 2026-03-06T21:17:28.250304+0000 mgr.vm03.uwuzgl (mgr.14199) 49 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: audit 2026-03-06T21:17:28.250304+0000 mgr.vm03.uwuzgl (mgr.14199) 49 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: audit 2026-03-06T21:17:28.251381+0000 mgr.vm03.uwuzgl (mgr.14199) 50 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm03.local:3000"}]: dispatch 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: audit 2026-03-06T21:17:28.251381+0000 mgr.vm03.uwuzgl (mgr.14199) 50 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm03.local:3000"}]: dispatch 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: audit 2026-03-06T21:17:28.263647+0000 mgr.vm03.uwuzgl (mgr.14199) 51 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: audit 2026-03-06T21:17:28.263647+0000 mgr.vm03.uwuzgl (mgr.14199) 51 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: audit 2026-03-06T21:17:28.264877+0000 mgr.vm03.uwuzgl (mgr.14199) 52 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm03.local:9093"}]: dispatch 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: audit 2026-03-06T21:17:28.264877+0000 mgr.vm03.uwuzgl (mgr.14199) 52 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm03.local:9093"}]: dispatch 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: audit 2026-03-06T21:17:28.276248+0000 mgr.vm03.uwuzgl (mgr.14199) 53 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: audit 2026-03-06T21:17:28.276248+0000 mgr.vm03.uwuzgl (mgr.14199) 53 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: audit 2026-03-06T21:17:28.277291+0000 mgr.vm03.uwuzgl (mgr.14199) 54 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm03.local:9095"}]: dispatch 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: audit 2026-03-06T21:17:28.277291+0000 mgr.vm03.uwuzgl (mgr.14199) 54 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm03.local:9095"}]: dispatch 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: audit 2026-03-06T21:17:29.164555+0000 mon.vm03 (mon.0) 329 : audit [DBG] from='client.? 192.168.123.103:0/3534283400' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:29.759 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:29 vm03 bash[17055]: audit 2026-03-06T21:17:29.164555+0000 mon.vm03 (mon.0) 329 : audit [DBG] from='client.? 192.168.123.103:0/3534283400' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:29.760 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records in 2026-03-06T22:17:29.760 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records out 2026-03-06T22:17:29.760 INFO:teuthology.orchestra.run.vm03.stderr:512 bytes copied, 0.000146424 s, 3.5 MB/s 2026-03-06T22:17:29.760 DEBUG:teuthology.orchestra.run.vm03:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-06T22:17:29.805 DEBUG:teuthology.orchestra.run.vm03:> stat /dev/vde 2026-03-06T22:17:29.852 INFO:teuthology.orchestra.run.vm03.stdout: File: /dev/vde 2026-03-06T22:17:29.852 INFO:teuthology.orchestra.run.vm03.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-06T22:17:29.852 INFO:teuthology.orchestra.run.vm03.stdout:Device: 5h/5d Inode: 27 Links: 1 Device type: fe,40 2026-03-06T22:17:29.852 INFO:teuthology.orchestra.run.vm03.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T22:17:29.852 INFO:teuthology.orchestra.run.vm03.stdout:Access: 2026-03-06 22:11:24.510585339 +0100 2026-03-06T22:17:29.852 INFO:teuthology.orchestra.run.vm03.stdout:Modify: 2026-03-06 22:11:22.726585339 +0100 2026-03-06T22:17:29.852 INFO:teuthology.orchestra.run.vm03.stdout:Change: 2026-03-06 22:11:22.726585339 +0100 2026-03-06T22:17:29.852 INFO:teuthology.orchestra.run.vm03.stdout: Birth: - 2026-03-06T22:17:29.852 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-06T22:17:29.899 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records in 2026-03-06T22:17:29.899 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records out 2026-03-06T22:17:29.899 INFO:teuthology.orchestra.run.vm03.stderr:512 bytes copied, 0.000128109 s, 4.0 MB/s 2026-03-06T22:17:29.899 DEBUG:teuthology.orchestra.run.vm03:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-06T22:17:29.944 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T22:17:29.944 DEBUG:teuthology.orchestra.run.vm08:> dd if=/scratch_devs of=/dev/stdout 2026-03-06T22:17:29.947 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T22:17:29.948 DEBUG:teuthology.orchestra.run.vm08:> ls /dev/[sv]d? 2026-03-06T22:17:29.991 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vda 2026-03-06T22:17:29.991 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdb 2026-03-06T22:17:29.991 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdc 2026-03-06T22:17:29.991 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdd 2026-03-06T22:17:29.991 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vde 2026-03-06T22:17:29.991 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-06T22:17:29.991 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-06T22:17:29.991 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdb 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: cephadm 2026-03-06T21:17:27.880927+0000 mgr.vm03.uwuzgl (mgr.14199) 46 : cephadm [INF] Reconfiguring ceph-exporter.vm08 (monmap changed)... 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: cephadm 2026-03-06T21:17:27.880927+0000 mgr.vm03.uwuzgl (mgr.14199) 46 : cephadm [INF] Reconfiguring ceph-exporter.vm08 (monmap changed)... 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: cephadm 2026-03-06T21:17:27.881958+0000 mgr.vm03.uwuzgl (mgr.14199) 47 : cephadm [INF] Reconfiguring daemon ceph-exporter.vm08 on vm08 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: cephadm 2026-03-06T21:17:27.881958+0000 mgr.vm03.uwuzgl (mgr.14199) 47 : cephadm [INF] Reconfiguring daemon ceph-exporter.vm08 on vm08 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: cluster 2026-03-06T21:17:28.022843+0000 mgr.vm03.uwuzgl (mgr.14199) 48 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: cluster 2026-03-06T21:17:28.022843+0000 mgr.vm03.uwuzgl (mgr.14199) 48 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: audit 2026-03-06T21:17:28.250304+0000 mgr.vm03.uwuzgl (mgr.14199) 49 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: audit 2026-03-06T21:17:28.250304+0000 mgr.vm03.uwuzgl (mgr.14199) 49 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: audit 2026-03-06T21:17:28.251381+0000 mgr.vm03.uwuzgl (mgr.14199) 50 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm03.local:3000"}]: dispatch 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: audit 2026-03-06T21:17:28.251381+0000 mgr.vm03.uwuzgl (mgr.14199) 50 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm03.local:3000"}]: dispatch 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: audit 2026-03-06T21:17:28.263647+0000 mgr.vm03.uwuzgl (mgr.14199) 51 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: audit 2026-03-06T21:17:28.263647+0000 mgr.vm03.uwuzgl (mgr.14199) 51 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: audit 2026-03-06T21:17:28.264877+0000 mgr.vm03.uwuzgl (mgr.14199) 52 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm03.local:9093"}]: dispatch 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: audit 2026-03-06T21:17:28.264877+0000 mgr.vm03.uwuzgl (mgr.14199) 52 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm03.local:9093"}]: dispatch 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: audit 2026-03-06T21:17:28.276248+0000 mgr.vm03.uwuzgl (mgr.14199) 53 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: audit 2026-03-06T21:17:28.276248+0000 mgr.vm03.uwuzgl (mgr.14199) 53 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: audit 2026-03-06T21:17:28.277291+0000 mgr.vm03.uwuzgl (mgr.14199) 54 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm03.local:9095"}]: dispatch 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: audit 2026-03-06T21:17:28.277291+0000 mgr.vm03.uwuzgl (mgr.14199) 54 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm03.local:9095"}]: dispatch 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: audit 2026-03-06T21:17:29.164555+0000 mon.vm03 (mon.0) 329 : audit [DBG] from='client.? 192.168.123.103:0/3534283400' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:29 vm08 bash[20828]: audit 2026-03-06T21:17:29.164555+0000 mon.vm03 (mon.0) 329 : audit [DBG] from='client.? 192.168.123.103:0/3534283400' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:30.004 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdb 2026-03-06T22:17:30.004 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-06T22:17:30.004 INFO:teuthology.orchestra.run.vm08.stdout:Device: 5h/5d Inode: 24 Links: 1 Device type: fe,10 2026-03-06T22:17:30.004 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T22:17:30.004 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-06 22:11:48.918875381 +0100 2026-03-06T22:17:30.004 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-06 22:11:47.690875381 +0100 2026-03-06T22:17:30.004 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-06 22:11:47.690875381 +0100 2026-03-06T22:17:30.004 INFO:teuthology.orchestra.run.vm08.stdout: Birth: - 2026-03-06T22:17:30.004 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-06T22:17:30.054 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-06T22:17:30.054 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-06T22:17:30.054 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000142467 s, 3.6 MB/s 2026-03-06T22:17:30.055 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-06T22:17:30.099 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdc 2026-03-06T22:17:30.143 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdc 2026-03-06T22:17:30.143 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-06T22:17:30.143 INFO:teuthology.orchestra.run.vm08.stdout:Device: 5h/5d Inode: 25 Links: 1 Device type: fe,20 2026-03-06T22:17:30.143 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T22:17:30.143 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-06 22:11:48.934875381 +0100 2026-03-06T22:17:30.143 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-06 22:11:47.694875381 +0100 2026-03-06T22:17:30.143 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-06 22:11:47.694875381 +0100 2026-03-06T22:17:30.143 INFO:teuthology.orchestra.run.vm08.stdout: Birth: - 2026-03-06T22:17:30.143 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-06T22:17:30.190 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-06T22:17:30.190 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-06T22:17:30.190 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000123883 s, 4.1 MB/s 2026-03-06T22:17:30.191 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-06T22:17:30.238 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdd 2026-03-06T22:17:30.283 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdd 2026-03-06T22:17:30.283 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-06T22:17:30.283 INFO:teuthology.orchestra.run.vm08.stdout:Device: 5h/5d Inode: 26 Links: 1 Device type: fe,30 2026-03-06T22:17:30.284 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T22:17:30.284 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-06 22:11:48.918875381 +0100 2026-03-06T22:17:30.284 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-06 22:11:47.702875381 +0100 2026-03-06T22:17:30.284 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-06 22:11:47.702875381 +0100 2026-03-06T22:17:30.284 INFO:teuthology.orchestra.run.vm08.stdout: Birth: - 2026-03-06T22:17:30.284 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-06T22:17:30.331 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-06T22:17:30.331 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-06T22:17:30.331 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000128431 s, 4.0 MB/s 2026-03-06T22:17:30.331 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-06T22:17:30.377 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vde 2026-03-06T22:17:30.424 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vde 2026-03-06T22:17:30.424 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-06T22:17:30.424 INFO:teuthology.orchestra.run.vm08.stdout:Device: 5h/5d Inode: 27 Links: 1 Device type: fe,40 2026-03-06T22:17:30.424 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T22:17:30.424 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-06 22:11:48.934875381 +0100 2026-03-06T22:17:30.424 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-06 22:11:47.690875381 +0100 2026-03-06T22:17:30.424 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-06 22:11:47.690875381 +0100 2026-03-06T22:17:30.424 INFO:teuthology.orchestra.run.vm08.stdout: Birth: - 2026-03-06T22:17:30.424 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-06T22:17:30.471 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-06T22:17:30.471 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-06T22:17:30.471 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000120947 s, 4.2 MB/s 2026-03-06T22:17:30.471 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-06T22:17:30.516 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph orch apply osd --all-available-devices 2026-03-06T22:17:31.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:31 vm03 bash[17055]: cluster 2026-03-06T21:17:30.022993+0000 mgr.vm03.uwuzgl (mgr.14199) 55 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:31.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:31 vm03 bash[17055]: cluster 2026-03-06T21:17:30.022993+0000 mgr.vm03.uwuzgl (mgr.14199) 55 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:32.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:31 vm08 bash[20828]: cluster 2026-03-06T21:17:30.022993+0000 mgr.vm03.uwuzgl (mgr.14199) 55 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:32.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:31 vm08 bash[20828]: cluster 2026-03-06T21:17:30.022993+0000 mgr.vm03.uwuzgl (mgr.14199) 55 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:33.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:33 vm03 bash[17055]: cluster 2026-03-06T21:17:32.023159+0000 mgr.vm03.uwuzgl (mgr.14199) 56 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:33.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:33 vm03 bash[17055]: cluster 2026-03-06T21:17:32.023159+0000 mgr.vm03.uwuzgl (mgr.14199) 56 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:33.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:33 vm03 bash[17055]: audit 2026-03-06T21:17:33.456734+0000 mon.vm03 (mon.0) 330 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:33.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:33 vm03 bash[17055]: audit 2026-03-06T21:17:33.456734+0000 mon.vm03 (mon.0) 330 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:33.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:33 vm03 bash[17055]: audit 2026-03-06T21:17:33.482324+0000 mon.vm03 (mon.0) 331 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:33.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:33 vm03 bash[17055]: audit 2026-03-06T21:17:33.482324+0000 mon.vm03 (mon.0) 331 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:33.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:33 vm03 bash[17055]: audit 2026-03-06T21:17:33.486271+0000 mon.vm03 (mon.0) 332 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:33.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:33 vm03 bash[17055]: audit 2026-03-06T21:17:33.486271+0000 mon.vm03 (mon.0) 332 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:33.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:33 vm03 bash[17055]: audit 2026-03-06T21:17:33.505390+0000 mon.vm03 (mon.0) 333 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:33.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:33 vm03 bash[17055]: audit 2026-03-06T21:17:33.505390+0000 mon.vm03 (mon.0) 333 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:33.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:33 vm03 bash[17055]: audit 2026-03-06T21:17:33.505953+0000 mon.vm03 (mon.0) 334 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:33.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:33 vm03 bash[17055]: audit 2026-03-06T21:17:33.505953+0000 mon.vm03 (mon.0) 334 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:33.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:33 vm03 bash[17055]: audit 2026-03-06T21:17:33.506393+0000 mon.vm03 (mon.0) 335 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:17:33.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:33 vm03 bash[17055]: audit 2026-03-06T21:17:33.506393+0000 mon.vm03 (mon.0) 335 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:17:33.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:33 vm03 bash[17055]: audit 2026-03-06T21:17:33.513171+0000 mon.vm03 (mon.0) 336 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:33.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:33 vm03 bash[17055]: audit 2026-03-06T21:17:33.513171+0000 mon.vm03 (mon.0) 336 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:33 vm08 bash[20828]: cluster 2026-03-06T21:17:32.023159+0000 mgr.vm03.uwuzgl (mgr.14199) 56 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:33 vm08 bash[20828]: cluster 2026-03-06T21:17:32.023159+0000 mgr.vm03.uwuzgl (mgr.14199) 56 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:33 vm08 bash[20828]: audit 2026-03-06T21:17:33.456734+0000 mon.vm03 (mon.0) 330 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:33 vm08 bash[20828]: audit 2026-03-06T21:17:33.456734+0000 mon.vm03 (mon.0) 330 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:33 vm08 bash[20828]: audit 2026-03-06T21:17:33.482324+0000 mon.vm03 (mon.0) 331 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:33 vm08 bash[20828]: audit 2026-03-06T21:17:33.482324+0000 mon.vm03 (mon.0) 331 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:33 vm08 bash[20828]: audit 2026-03-06T21:17:33.486271+0000 mon.vm03 (mon.0) 332 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:33 vm08 bash[20828]: audit 2026-03-06T21:17:33.486271+0000 mon.vm03 (mon.0) 332 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:33 vm08 bash[20828]: audit 2026-03-06T21:17:33.505390+0000 mon.vm03 (mon.0) 333 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:33 vm08 bash[20828]: audit 2026-03-06T21:17:33.505390+0000 mon.vm03 (mon.0) 333 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:33 vm08 bash[20828]: audit 2026-03-06T21:17:33.505953+0000 mon.vm03 (mon.0) 334 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:33 vm08 bash[20828]: audit 2026-03-06T21:17:33.505953+0000 mon.vm03 (mon.0) 334 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:33 vm08 bash[20828]: audit 2026-03-06T21:17:33.506393+0000 mon.vm03 (mon.0) 335 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:17:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:33 vm08 bash[20828]: audit 2026-03-06T21:17:33.506393+0000 mon.vm03 (mon.0) 335 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:17:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:33 vm08 bash[20828]: audit 2026-03-06T21:17:33.513171+0000 mon.vm03 (mon.0) 336 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:33 vm08 bash[20828]: audit 2026-03-06T21:17:33.513171+0000 mon.vm03 (mon.0) 336 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:34.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:34 vm03 bash[17055]: cluster 2026-03-06T21:17:34.023309+0000 mgr.vm03.uwuzgl (mgr.14199) 57 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:34.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:34 vm03 bash[17055]: cluster 2026-03-06T21:17:34.023309+0000 mgr.vm03.uwuzgl (mgr.14199) 57 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:35.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:34 vm08 bash[20828]: cluster 2026-03-06T21:17:34.023309+0000 mgr.vm03.uwuzgl (mgr.14199) 57 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:35.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:34 vm08 bash[20828]: cluster 2026-03-06T21:17:34.023309+0000 mgr.vm03.uwuzgl (mgr.14199) 57 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:35.108 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm08/config 2026-03-06T22:17:35.468 INFO:teuthology.orchestra.run.vm08.stdout:Scheduled osd.all-available-devices update... 2026-03-06T22:17:35.544 INFO:tasks.cephadm:Waiting for 8 OSDs to come up... 2026-03-06T22:17:35.544 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd stat -f json 2026-03-06T22:17:36.470 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:36 vm08 bash[20828]: audit 2026-03-06T21:17:35.456491+0000 mgr.vm03.uwuzgl (mgr.14199) 58 : audit [DBG] from='client.24101 -' entity='client.admin' cmd=[{"prefix": "orch apply osd", "all_available_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:17:36.471 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:36 vm08 bash[20828]: audit 2026-03-06T21:17:35.456491+0000 mgr.vm03.uwuzgl (mgr.14199) 58 : audit [DBG] from='client.24101 -' entity='client.admin' cmd=[{"prefix": "orch apply osd", "all_available_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:17:36.471 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:36 vm08 bash[20828]: cephadm 2026-03-06T21:17:35.457432+0000 mgr.vm03.uwuzgl (mgr.14199) 59 : cephadm [INF] Marking host: vm03 for OSDSpec preview refresh. 2026-03-06T22:17:36.471 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:36 vm08 bash[20828]: cephadm 2026-03-06T21:17:35.457432+0000 mgr.vm03.uwuzgl (mgr.14199) 59 : cephadm [INF] Marking host: vm03 for OSDSpec preview refresh. 2026-03-06T22:17:36.471 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:36 vm08 bash[20828]: cephadm 2026-03-06T21:17:35.457453+0000 mgr.vm03.uwuzgl (mgr.14199) 60 : cephadm [INF] Marking host: vm08 for OSDSpec preview refresh. 2026-03-06T22:17:36.471 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:36 vm08 bash[20828]: cephadm 2026-03-06T21:17:35.457453+0000 mgr.vm03.uwuzgl (mgr.14199) 60 : cephadm [INF] Marking host: vm08 for OSDSpec preview refresh. 2026-03-06T22:17:36.471 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:36 vm08 bash[20828]: cephadm 2026-03-06T21:17:35.457618+0000 mgr.vm03.uwuzgl (mgr.14199) 61 : cephadm [INF] Saving service osd.all-available-devices spec with placement * 2026-03-06T22:17:36.471 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:36 vm08 bash[20828]: cephadm 2026-03-06T21:17:35.457618+0000 mgr.vm03.uwuzgl (mgr.14199) 61 : cephadm [INF] Saving service osd.all-available-devices spec with placement * 2026-03-06T22:17:36.471 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:36 vm08 bash[20828]: audit 2026-03-06T21:17:35.462028+0000 mon.vm03 (mon.0) 337 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:36.471 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:36 vm08 bash[20828]: audit 2026-03-06T21:17:35.462028+0000 mon.vm03 (mon.0) 337 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:36.471 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:36 vm08 bash[20828]: audit 2026-03-06T21:17:35.462663+0000 mon.vm03 (mon.0) 338 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:17:36.471 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:36 vm08 bash[20828]: audit 2026-03-06T21:17:35.462663+0000 mon.vm03 (mon.0) 338 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:17:36.471 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:36 vm08 bash[20828]: audit 2026-03-06T21:17:36.056241+0000 mon.vm03 (mon.0) 339 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:17:36.471 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:36 vm08 bash[20828]: audit 2026-03-06T21:17:36.056241+0000 mon.vm03 (mon.0) 339 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:17:36.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:36 vm03 bash[17055]: audit 2026-03-06T21:17:35.456491+0000 mgr.vm03.uwuzgl (mgr.14199) 58 : audit [DBG] from='client.24101 -' entity='client.admin' cmd=[{"prefix": "orch apply osd", "all_available_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:17:36.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:36 vm03 bash[17055]: audit 2026-03-06T21:17:35.456491+0000 mgr.vm03.uwuzgl (mgr.14199) 58 : audit [DBG] from='client.24101 -' entity='client.admin' cmd=[{"prefix": "orch apply osd", "all_available_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:17:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:36 vm03 bash[17055]: cephadm 2026-03-06T21:17:35.457432+0000 mgr.vm03.uwuzgl (mgr.14199) 59 : cephadm [INF] Marking host: vm03 for OSDSpec preview refresh. 2026-03-06T22:17:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:36 vm03 bash[17055]: cephadm 2026-03-06T21:17:35.457432+0000 mgr.vm03.uwuzgl (mgr.14199) 59 : cephadm [INF] Marking host: vm03 for OSDSpec preview refresh. 2026-03-06T22:17:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:36 vm03 bash[17055]: cephadm 2026-03-06T21:17:35.457453+0000 mgr.vm03.uwuzgl (mgr.14199) 60 : cephadm [INF] Marking host: vm08 for OSDSpec preview refresh. 2026-03-06T22:17:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:36 vm03 bash[17055]: cephadm 2026-03-06T21:17:35.457453+0000 mgr.vm03.uwuzgl (mgr.14199) 60 : cephadm [INF] Marking host: vm08 for OSDSpec preview refresh. 2026-03-06T22:17:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:36 vm03 bash[17055]: cephadm 2026-03-06T21:17:35.457618+0000 mgr.vm03.uwuzgl (mgr.14199) 61 : cephadm [INF] Saving service osd.all-available-devices spec with placement * 2026-03-06T22:17:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:36 vm03 bash[17055]: cephadm 2026-03-06T21:17:35.457618+0000 mgr.vm03.uwuzgl (mgr.14199) 61 : cephadm [INF] Saving service osd.all-available-devices spec with placement * 2026-03-06T22:17:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:36 vm03 bash[17055]: audit 2026-03-06T21:17:35.462028+0000 mon.vm03 (mon.0) 337 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:36 vm03 bash[17055]: audit 2026-03-06T21:17:35.462028+0000 mon.vm03 (mon.0) 337 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:36 vm03 bash[17055]: audit 2026-03-06T21:17:35.462663+0000 mon.vm03 (mon.0) 338 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:17:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:36 vm03 bash[17055]: audit 2026-03-06T21:17:35.462663+0000 mon.vm03 (mon.0) 338 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:17:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:36 vm03 bash[17055]: audit 2026-03-06T21:17:36.056241+0000 mon.vm03 (mon.0) 339 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:17:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:36 vm03 bash[17055]: audit 2026-03-06T21:17:36.056241+0000 mon.vm03 (mon.0) 339 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:17:37.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:37 vm08 bash[20828]: cluster 2026-03-06T21:17:36.023478+0000 mgr.vm03.uwuzgl (mgr.14199) 62 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:37.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:37 vm08 bash[20828]: cluster 2026-03-06T21:17:36.023478+0000 mgr.vm03.uwuzgl (mgr.14199) 62 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:37.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:37 vm03 bash[17055]: cluster 2026-03-06T21:17:36.023478+0000 mgr.vm03.uwuzgl (mgr.14199) 62 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:37.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:37 vm03 bash[17055]: cluster 2026-03-06T21:17:36.023478+0000 mgr.vm03.uwuzgl (mgr.14199) 62 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:39.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:38 vm08 bash[20828]: cluster 2026-03-06T21:17:38.023683+0000 mgr.vm03.uwuzgl (mgr.14199) 63 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:39.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:38 vm08 bash[20828]: cluster 2026-03-06T21:17:38.023683+0000 mgr.vm03.uwuzgl (mgr.14199) 63 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:39.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:38 vm03 bash[17055]: cluster 2026-03-06T21:17:38.023683+0000 mgr.vm03.uwuzgl (mgr.14199) 63 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:39.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:38 vm03 bash[17055]: cluster 2026-03-06T21:17:38.023683+0000 mgr.vm03.uwuzgl (mgr.14199) 63 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:40.241 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:17:40.687 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:17:40.776 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":5,"num_osds":0,"num_up_osds":0,"osd_up_since":0,"num_in_osds":0,"osd_in_since":0,"num_remapped_pgs":0} 2026-03-06T22:17:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: cluster 2026-03-06T21:17:40.023857+0000 mgr.vm03.uwuzgl (mgr.14199) 64 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: cluster 2026-03-06T21:17:40.023857+0000 mgr.vm03.uwuzgl (mgr.14199) 64 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.487029+0000 mon.vm03 (mon.0) 340 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.487029+0000 mon.vm03 (mon.0) 340 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.494820+0000 mon.vm03 (mon.0) 341 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.494820+0000 mon.vm03 (mon.0) 341 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.499550+0000 mon.vm03 (mon.0) 342 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.499550+0000 mon.vm03 (mon.0) 342 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.504970+0000 mon.vm03 (mon.0) 343 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.504970+0000 mon.vm03 (mon.0) 343 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.681630+0000 mon.vm03 (mon.0) 344 : audit [DBG] from='client.? 192.168.123.103:0/2434994782' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.681630+0000 mon.vm03 (mon.0) 344 : audit [DBG] from='client.? 192.168.123.103:0/2434994782' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.893494+0000 mon.vm03 (mon.0) 345 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.893494+0000 mon.vm03 (mon.0) 345 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.897602+0000 mon.vm03 (mon.0) 346 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.897602+0000 mon.vm03 (mon.0) 346 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.901186+0000 mon.vm03 (mon.0) 347 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.901186+0000 mon.vm03 (mon.0) 347 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.904764+0000 mon.vm03 (mon.0) 348 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.904764+0000 mon.vm03 (mon.0) 348 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.905415+0000 mon.vm03 (mon.0) 349 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.905415+0000 mon.vm03 (mon.0) 349 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.905941+0000 mon.vm03 (mon.0) 350 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.905941+0000 mon.vm03 (mon.0) 350 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.908725+0000 mon.vm03 (mon.0) 351 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.908725+0000 mon.vm03 (mon.0) 351 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.909894+0000 mon.vm03 (mon.0) 352 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.909894+0000 mon.vm03 (mon.0) 352 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.911692+0000 mon.vm03 (mon.0) 353 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.911692+0000 mon.vm03 (mon.0) 353 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.912093+0000 mon.vm03 (mon.0) 354 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.912093+0000 mon.vm03 (mon.0) 354 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.913961+0000 mon.vm03 (mon.0) 355 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.913961+0000 mon.vm03 (mon.0) 355 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.914352+0000 mon.vm03 (mon.0) 356 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:41.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:41 vm08 bash[20828]: audit 2026-03-06T21:17:40.914352+0000 mon.vm03 (mon.0) 356 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:41.777 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd stat -f json 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: cluster 2026-03-06T21:17:40.023857+0000 mgr.vm03.uwuzgl (mgr.14199) 64 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: cluster 2026-03-06T21:17:40.023857+0000 mgr.vm03.uwuzgl (mgr.14199) 64 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.487029+0000 mon.vm03 (mon.0) 340 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.487029+0000 mon.vm03 (mon.0) 340 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.494820+0000 mon.vm03 (mon.0) 341 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.494820+0000 mon.vm03 (mon.0) 341 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.499550+0000 mon.vm03 (mon.0) 342 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.499550+0000 mon.vm03 (mon.0) 342 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.504970+0000 mon.vm03 (mon.0) 343 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.504970+0000 mon.vm03 (mon.0) 343 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.681630+0000 mon.vm03 (mon.0) 344 : audit [DBG] from='client.? 192.168.123.103:0/2434994782' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.681630+0000 mon.vm03 (mon.0) 344 : audit [DBG] from='client.? 192.168.123.103:0/2434994782' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.893494+0000 mon.vm03 (mon.0) 345 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.893494+0000 mon.vm03 (mon.0) 345 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.897602+0000 mon.vm03 (mon.0) 346 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.897602+0000 mon.vm03 (mon.0) 346 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.901186+0000 mon.vm03 (mon.0) 347 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.901186+0000 mon.vm03 (mon.0) 347 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.904764+0000 mon.vm03 (mon.0) 348 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.904764+0000 mon.vm03 (mon.0) 348 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.905415+0000 mon.vm03 (mon.0) 349 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.905415+0000 mon.vm03 (mon.0) 349 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.905941+0000 mon.vm03 (mon.0) 350 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.905941+0000 mon.vm03 (mon.0) 350 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.908725+0000 mon.vm03 (mon.0) 351 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.908725+0000 mon.vm03 (mon.0) 351 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.909894+0000 mon.vm03 (mon.0) 352 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.909894+0000 mon.vm03 (mon.0) 352 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:17:41.783 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.911692+0000 mon.vm03 (mon.0) 353 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T22:17:41.784 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.911692+0000 mon.vm03 (mon.0) 353 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T22:17:41.784 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.912093+0000 mon.vm03 (mon.0) 354 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:41.784 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.912093+0000 mon.vm03 (mon.0) 354 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:41.784 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.913961+0000 mon.vm03 (mon.0) 355 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T22:17:41.784 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.913961+0000 mon.vm03 (mon.0) 355 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T22:17:41.784 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.914352+0000 mon.vm03 (mon.0) 356 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:41.784 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:41 vm03 bash[17055]: audit 2026-03-06T21:17:40.914352+0000 mon.vm03 (mon.0) 356 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:17:43.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:43 vm08 bash[20828]: cluster 2026-03-06T21:17:42.024045+0000 mgr.vm03.uwuzgl (mgr.14199) 65 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:43.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:43 vm08 bash[20828]: cluster 2026-03-06T21:17:42.024045+0000 mgr.vm03.uwuzgl (mgr.14199) 65 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:43.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:43 vm03 bash[17055]: cluster 2026-03-06T21:17:42.024045+0000 mgr.vm03.uwuzgl (mgr.14199) 65 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:43.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:43 vm03 bash[17055]: cluster 2026-03-06T21:17:42.024045+0000 mgr.vm03.uwuzgl (mgr.14199) 65 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:45.692 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:17:45.758 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:45 vm08 bash[20828]: cluster 2026-03-06T21:17:44.024258+0000 mgr.vm03.uwuzgl (mgr.14199) 66 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:45.759 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:45 vm08 bash[20828]: cluster 2026-03-06T21:17:44.024258+0000 mgr.vm03.uwuzgl (mgr.14199) 66 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:45.760 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:45 vm03 bash[17055]: cluster 2026-03-06T21:17:44.024258+0000 mgr.vm03.uwuzgl (mgr.14199) 66 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:45.760 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:45 vm03 bash[17055]: cluster 2026-03-06T21:17:44.024258+0000 mgr.vm03.uwuzgl (mgr.14199) 66 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:46.142 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:17:46.234 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":5,"num_osds":0,"num_up_osds":0,"osd_up_since":0,"num_in_osds":0,"osd_in_since":0,"num_remapped_pgs":0} 2026-03-06T22:17:46.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:46 vm03 bash[17055]: audit 2026-03-06T21:17:46.135325+0000 mon.vm08 (mon.1) 2 : audit [DBG] from='client.? 192.168.123.103:0/3897606211' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:17:46.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:46 vm03 bash[17055]: audit 2026-03-06T21:17:46.135325+0000 mon.vm08 (mon.1) 2 : audit [DBG] from='client.? 192.168.123.103:0/3897606211' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:17:47.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:46 vm08 bash[20828]: audit 2026-03-06T21:17:46.135325+0000 mon.vm08 (mon.1) 2 : audit [DBG] from='client.? 192.168.123.103:0/3897606211' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:17:47.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:46 vm08 bash[20828]: audit 2026-03-06T21:17:46.135325+0000 mon.vm08 (mon.1) 2 : audit [DBG] from='client.? 192.168.123.103:0/3897606211' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:17:47.236 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd stat -f json 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: cluster 2026-03-06T21:17:46.057025+0000 mgr.vm03.uwuzgl (mgr.14199) 67 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: cluster 2026-03-06T21:17:46.057025+0000 mgr.vm03.uwuzgl (mgr.14199) 67 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: audit 2026-03-06T21:17:47.055995+0000 mon.vm03 (mon.0) 357 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "198867f4-f98d-4ee8-a08e-17f3733bec4d"}]: dispatch 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: audit 2026-03-06T21:17:47.055995+0000 mon.vm03 (mon.0) 357 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "198867f4-f98d-4ee8-a08e-17f3733bec4d"}]: dispatch 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: audit 2026-03-06T21:17:47.058866+0000 mon.vm03 (mon.0) 358 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "198867f4-f98d-4ee8-a08e-17f3733bec4d"}]': finished 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: audit 2026-03-06T21:17:47.058866+0000 mon.vm03 (mon.0) 358 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "198867f4-f98d-4ee8-a08e-17f3733bec4d"}]': finished 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: audit 2026-03-06T21:17:47.059991+0000 mon.vm08 (mon.1) 3 : audit [INF] from='client.? 192.168.123.108:0/1231215453' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "198867f4-f98d-4ee8-a08e-17f3733bec4d"}]: dispatch 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: audit 2026-03-06T21:17:47.059991+0000 mon.vm08 (mon.1) 3 : audit [INF] from='client.? 192.168.123.108:0/1231215453' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "198867f4-f98d-4ee8-a08e-17f3733bec4d"}]: dispatch 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: cluster 2026-03-06T21:17:47.060779+0000 mon.vm03 (mon.0) 359 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: cluster 2026-03-06T21:17:47.060779+0000 mon.vm03 (mon.0) 359 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: audit 2026-03-06T21:17:47.060865+0000 mon.vm03 (mon.0) 360 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: audit 2026-03-06T21:17:47.060865+0000 mon.vm03 (mon.0) 360 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: audit 2026-03-06T21:17:47.122456+0000 mon.vm03 (mon.0) 361 : audit [INF] from='client.? 192.168.123.103:0/2001668021' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "b3b4d223-7698-4548-bde2-0d448c9b3714"}]: dispatch 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: audit 2026-03-06T21:17:47.122456+0000 mon.vm03 (mon.0) 361 : audit [INF] from='client.? 192.168.123.103:0/2001668021' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "b3b4d223-7698-4548-bde2-0d448c9b3714"}]: dispatch 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: audit 2026-03-06T21:17:47.124872+0000 mon.vm03 (mon.0) 362 : audit [INF] from='client.? 192.168.123.103:0/2001668021' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "b3b4d223-7698-4548-bde2-0d448c9b3714"}]': finished 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: audit 2026-03-06T21:17:47.124872+0000 mon.vm03 (mon.0) 362 : audit [INF] from='client.? 192.168.123.103:0/2001668021' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "b3b4d223-7698-4548-bde2-0d448c9b3714"}]': finished 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: cluster 2026-03-06T21:17:47.126881+0000 mon.vm03 (mon.0) 363 : cluster [DBG] osdmap e7: 2 total, 0 up, 2 in 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: cluster 2026-03-06T21:17:47.126881+0000 mon.vm03 (mon.0) 363 : cluster [DBG] osdmap e7: 2 total, 0 up, 2 in 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: audit 2026-03-06T21:17:47.126980+0000 mon.vm03 (mon.0) 364 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: audit 2026-03-06T21:17:47.126980+0000 mon.vm03 (mon.0) 364 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: audit 2026-03-06T21:17:47.127024+0000 mon.vm03 (mon.0) 365 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:47.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:47 vm03 bash[17055]: audit 2026-03-06T21:17:47.127024+0000 mon.vm03 (mon.0) 365 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:48.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: cluster 2026-03-06T21:17:46.057025+0000 mgr.vm03.uwuzgl (mgr.14199) 67 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:48.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: cluster 2026-03-06T21:17:46.057025+0000 mgr.vm03.uwuzgl (mgr.14199) 67 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: audit 2026-03-06T21:17:47.055995+0000 mon.vm03 (mon.0) 357 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "198867f4-f98d-4ee8-a08e-17f3733bec4d"}]: dispatch 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: audit 2026-03-06T21:17:47.055995+0000 mon.vm03 (mon.0) 357 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "198867f4-f98d-4ee8-a08e-17f3733bec4d"}]: dispatch 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: audit 2026-03-06T21:17:47.058866+0000 mon.vm03 (mon.0) 358 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "198867f4-f98d-4ee8-a08e-17f3733bec4d"}]': finished 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: audit 2026-03-06T21:17:47.058866+0000 mon.vm03 (mon.0) 358 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "198867f4-f98d-4ee8-a08e-17f3733bec4d"}]': finished 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: audit 2026-03-06T21:17:47.059991+0000 mon.vm08 (mon.1) 3 : audit [INF] from='client.? 192.168.123.108:0/1231215453' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "198867f4-f98d-4ee8-a08e-17f3733bec4d"}]: dispatch 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: audit 2026-03-06T21:17:47.059991+0000 mon.vm08 (mon.1) 3 : audit [INF] from='client.? 192.168.123.108:0/1231215453' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "198867f4-f98d-4ee8-a08e-17f3733bec4d"}]: dispatch 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: cluster 2026-03-06T21:17:47.060779+0000 mon.vm03 (mon.0) 359 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: cluster 2026-03-06T21:17:47.060779+0000 mon.vm03 (mon.0) 359 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: audit 2026-03-06T21:17:47.060865+0000 mon.vm03 (mon.0) 360 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: audit 2026-03-06T21:17:47.060865+0000 mon.vm03 (mon.0) 360 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: audit 2026-03-06T21:17:47.122456+0000 mon.vm03 (mon.0) 361 : audit [INF] from='client.? 192.168.123.103:0/2001668021' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "b3b4d223-7698-4548-bde2-0d448c9b3714"}]: dispatch 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: audit 2026-03-06T21:17:47.122456+0000 mon.vm03 (mon.0) 361 : audit [INF] from='client.? 192.168.123.103:0/2001668021' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "b3b4d223-7698-4548-bde2-0d448c9b3714"}]: dispatch 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: audit 2026-03-06T21:17:47.124872+0000 mon.vm03 (mon.0) 362 : audit [INF] from='client.? 192.168.123.103:0/2001668021' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "b3b4d223-7698-4548-bde2-0d448c9b3714"}]': finished 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: audit 2026-03-06T21:17:47.124872+0000 mon.vm03 (mon.0) 362 : audit [INF] from='client.? 192.168.123.103:0/2001668021' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "b3b4d223-7698-4548-bde2-0d448c9b3714"}]': finished 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: cluster 2026-03-06T21:17:47.126881+0000 mon.vm03 (mon.0) 363 : cluster [DBG] osdmap e7: 2 total, 0 up, 2 in 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: cluster 2026-03-06T21:17:47.126881+0000 mon.vm03 (mon.0) 363 : cluster [DBG] osdmap e7: 2 total, 0 up, 2 in 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: audit 2026-03-06T21:17:47.126980+0000 mon.vm03 (mon.0) 364 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: audit 2026-03-06T21:17:47.126980+0000 mon.vm03 (mon.0) 364 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: audit 2026-03-06T21:17:47.127024+0000 mon.vm03 (mon.0) 365 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:48.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:47 vm08 bash[20828]: audit 2026-03-06T21:17:47.127024+0000 mon.vm03 (mon.0) 365 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:48.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:48 vm03 bash[17055]: audit 2026-03-06T21:17:47.672305+0000 mon.vm08 (mon.1) 4 : audit [DBG] from='client.? 192.168.123.108:0/1388798953' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:48.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:48 vm03 bash[17055]: audit 2026-03-06T21:17:47.672305+0000 mon.vm08 (mon.1) 4 : audit [DBG] from='client.? 192.168.123.108:0/1388798953' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:48.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:48 vm03 bash[17055]: audit 2026-03-06T21:17:47.844414+0000 mon.vm03 (mon.0) 366 : audit [DBG] from='client.? 192.168.123.103:0/3216892498' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:48.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:48 vm03 bash[17055]: audit 2026-03-06T21:17:47.844414+0000 mon.vm03 (mon.0) 366 : audit [DBG] from='client.? 192.168.123.103:0/3216892498' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:49.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:48 vm08 bash[20828]: audit 2026-03-06T21:17:47.672305+0000 mon.vm08 (mon.1) 4 : audit [DBG] from='client.? 192.168.123.108:0/1388798953' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:49.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:48 vm08 bash[20828]: audit 2026-03-06T21:17:47.672305+0000 mon.vm08 (mon.1) 4 : audit [DBG] from='client.? 192.168.123.108:0/1388798953' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:49.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:48 vm08 bash[20828]: audit 2026-03-06T21:17:47.844414+0000 mon.vm03 (mon.0) 366 : audit [DBG] from='client.? 192.168.123.103:0/3216892498' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:49.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:48 vm08 bash[20828]: audit 2026-03-06T21:17:47.844414+0000 mon.vm03 (mon.0) 366 : audit [DBG] from='client.? 192.168.123.103:0/3216892498' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:49.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:49 vm03 bash[17055]: cluster 2026-03-06T21:17:48.057243+0000 mgr.vm03.uwuzgl (mgr.14199) 68 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:49.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:49 vm03 bash[17055]: cluster 2026-03-06T21:17:48.057243+0000 mgr.vm03.uwuzgl (mgr.14199) 68 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:50.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:49 vm08 bash[20828]: cluster 2026-03-06T21:17:48.057243+0000 mgr.vm03.uwuzgl (mgr.14199) 68 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:50.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:49 vm08 bash[20828]: cluster 2026-03-06T21:17:48.057243+0000 mgr.vm03.uwuzgl (mgr.14199) 68 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:50.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:50 vm03 bash[17055]: cluster 2026-03-06T21:17:50.057446+0000 mgr.vm03.uwuzgl (mgr.14199) 69 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:50.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:50 vm03 bash[17055]: cluster 2026-03-06T21:17:50.057446+0000 mgr.vm03.uwuzgl (mgr.14199) 69 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:51.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:50 vm08 bash[20828]: cluster 2026-03-06T21:17:50.057446+0000 mgr.vm03.uwuzgl (mgr.14199) 69 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:51.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:50 vm08 bash[20828]: cluster 2026-03-06T21:17:50.057446+0000 mgr.vm03.uwuzgl (mgr.14199) 69 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:51.842 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:51 vm03 bash[17055]: audit 2026-03-06T21:17:51.056606+0000 mon.vm03 (mon.0) 367 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:17:51.843 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:51 vm03 bash[17055]: audit 2026-03-06T21:17:51.056606+0000 mon.vm03 (mon.0) 367 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:17:51.843 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:51 vm03 bash[17055]: audit 2026-03-06T21:17:51.511005+0000 mon.vm03 (mon.0) 368 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a6b95b38-64b6-4e58-94c4-ae263e2dc1ff"}]: dispatch 2026-03-06T22:17:51.843 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:51 vm03 bash[17055]: audit 2026-03-06T21:17:51.511005+0000 mon.vm03 (mon.0) 368 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a6b95b38-64b6-4e58-94c4-ae263e2dc1ff"}]: dispatch 2026-03-06T22:17:51.843 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:51 vm03 bash[17055]: audit 2026-03-06T21:17:51.514140+0000 mon.vm08 (mon.1) 5 : audit [INF] from='client.? 192.168.123.108:0/1129615040' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a6b95b38-64b6-4e58-94c4-ae263e2dc1ff"}]: dispatch 2026-03-06T22:17:51.843 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:51 vm03 bash[17055]: audit 2026-03-06T21:17:51.514140+0000 mon.vm08 (mon.1) 5 : audit [INF] from='client.? 192.168.123.108:0/1129615040' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a6b95b38-64b6-4e58-94c4-ae263e2dc1ff"}]: dispatch 2026-03-06T22:17:51.843 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:51 vm03 bash[17055]: audit 2026-03-06T21:17:51.521655+0000 mon.vm03 (mon.0) 369 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a6b95b38-64b6-4e58-94c4-ae263e2dc1ff"}]': finished 2026-03-06T22:17:51.843 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:51 vm03 bash[17055]: audit 2026-03-06T21:17:51.521655+0000 mon.vm03 (mon.0) 369 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a6b95b38-64b6-4e58-94c4-ae263e2dc1ff"}]': finished 2026-03-06T22:17:51.843 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:51 vm03 bash[17055]: cluster 2026-03-06T21:17:51.524363+0000 mon.vm03 (mon.0) 370 : cluster [DBG] osdmap e8: 3 total, 0 up, 3 in 2026-03-06T22:17:51.843 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:51 vm03 bash[17055]: cluster 2026-03-06T21:17:51.524363+0000 mon.vm03 (mon.0) 370 : cluster [DBG] osdmap e8: 3 total, 0 up, 3 in 2026-03-06T22:17:51.843 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:51 vm03 bash[17055]: audit 2026-03-06T21:17:51.524546+0000 mon.vm03 (mon.0) 371 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:51.843 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:51 vm03 bash[17055]: audit 2026-03-06T21:17:51.524546+0000 mon.vm03 (mon.0) 371 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:51.843 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:51 vm03 bash[17055]: audit 2026-03-06T21:17:51.524592+0000 mon.vm03 (mon.0) 372 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:51.843 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:51 vm03 bash[17055]: audit 2026-03-06T21:17:51.524592+0000 mon.vm03 (mon.0) 372 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:51.843 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:51 vm03 bash[17055]: audit 2026-03-06T21:17:51.524623+0000 mon.vm03 (mon.0) 373 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:51.843 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:51 vm03 bash[17055]: audit 2026-03-06T21:17:51.524623+0000 mon.vm03 (mon.0) 373 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:52.000 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:17:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:51 vm08 bash[20828]: audit 2026-03-06T21:17:51.056606+0000 mon.vm03 (mon.0) 367 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:17:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:51 vm08 bash[20828]: audit 2026-03-06T21:17:51.056606+0000 mon.vm03 (mon.0) 367 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:17:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:51 vm08 bash[20828]: audit 2026-03-06T21:17:51.511005+0000 mon.vm03 (mon.0) 368 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a6b95b38-64b6-4e58-94c4-ae263e2dc1ff"}]: dispatch 2026-03-06T22:17:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:51 vm08 bash[20828]: audit 2026-03-06T21:17:51.511005+0000 mon.vm03 (mon.0) 368 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a6b95b38-64b6-4e58-94c4-ae263e2dc1ff"}]: dispatch 2026-03-06T22:17:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:51 vm08 bash[20828]: audit 2026-03-06T21:17:51.514140+0000 mon.vm08 (mon.1) 5 : audit [INF] from='client.? 192.168.123.108:0/1129615040' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a6b95b38-64b6-4e58-94c4-ae263e2dc1ff"}]: dispatch 2026-03-06T22:17:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:51 vm08 bash[20828]: audit 2026-03-06T21:17:51.514140+0000 mon.vm08 (mon.1) 5 : audit [INF] from='client.? 192.168.123.108:0/1129615040' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a6b95b38-64b6-4e58-94c4-ae263e2dc1ff"}]: dispatch 2026-03-06T22:17:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:51 vm08 bash[20828]: audit 2026-03-06T21:17:51.521655+0000 mon.vm03 (mon.0) 369 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a6b95b38-64b6-4e58-94c4-ae263e2dc1ff"}]': finished 2026-03-06T22:17:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:51 vm08 bash[20828]: audit 2026-03-06T21:17:51.521655+0000 mon.vm03 (mon.0) 369 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a6b95b38-64b6-4e58-94c4-ae263e2dc1ff"}]': finished 2026-03-06T22:17:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:51 vm08 bash[20828]: cluster 2026-03-06T21:17:51.524363+0000 mon.vm03 (mon.0) 370 : cluster [DBG] osdmap e8: 3 total, 0 up, 3 in 2026-03-06T22:17:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:51 vm08 bash[20828]: cluster 2026-03-06T21:17:51.524363+0000 mon.vm03 (mon.0) 370 : cluster [DBG] osdmap e8: 3 total, 0 up, 3 in 2026-03-06T22:17:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:51 vm08 bash[20828]: audit 2026-03-06T21:17:51.524546+0000 mon.vm03 (mon.0) 371 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:51 vm08 bash[20828]: audit 2026-03-06T21:17:51.524546+0000 mon.vm03 (mon.0) 371 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:51 vm08 bash[20828]: audit 2026-03-06T21:17:51.524592+0000 mon.vm03 (mon.0) 372 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:51 vm08 bash[20828]: audit 2026-03-06T21:17:51.524592+0000 mon.vm03 (mon.0) 372 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:51 vm08 bash[20828]: audit 2026-03-06T21:17:51.524623+0000 mon.vm03 (mon.0) 373 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:51 vm08 bash[20828]: audit 2026-03-06T21:17:51.524623+0000 mon.vm03 (mon.0) 373 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:52.412 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:17:52.468 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":9,"num_osds":4,"num_up_osds":0,"osd_up_since":0,"num_in_osds":4,"osd_in_since":1772831871,"num_remapped_pgs":0} 2026-03-06T22:17:52.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:51.710265+0000 mon.vm03 (mon.0) 374 : audit [INF] from='client.? 192.168.123.103:0/2076291687' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8bb1dfca-f66a-4fa1-86a3-bb356250604f"}]: dispatch 2026-03-06T22:17:52.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:51.710265+0000 mon.vm03 (mon.0) 374 : audit [INF] from='client.? 192.168.123.103:0/2076291687' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8bb1dfca-f66a-4fa1-86a3-bb356250604f"}]: dispatch 2026-03-06T22:17:52.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:51.713271+0000 mon.vm03 (mon.0) 375 : audit [INF] from='client.? 192.168.123.103:0/2076291687' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8bb1dfca-f66a-4fa1-86a3-bb356250604f"}]': finished 2026-03-06T22:17:52.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:51.713271+0000 mon.vm03 (mon.0) 375 : audit [INF] from='client.? 192.168.123.103:0/2076291687' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8bb1dfca-f66a-4fa1-86a3-bb356250604f"}]': finished 2026-03-06T22:17:52.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: cluster 2026-03-06T21:17:51.715557+0000 mon.vm03 (mon.0) 376 : cluster [DBG] osdmap e9: 4 total, 0 up, 4 in 2026-03-06T22:17:52.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: cluster 2026-03-06T21:17:51.715557+0000 mon.vm03 (mon.0) 376 : cluster [DBG] osdmap e9: 4 total, 0 up, 4 in 2026-03-06T22:17:52.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:51.715708+0000 mon.vm03 (mon.0) 377 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:52.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:51.715708+0000 mon.vm03 (mon.0) 377 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:52.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:51.715801+0000 mon.vm03 (mon.0) 378 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:52.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:51.715801+0000 mon.vm03 (mon.0) 378 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:52.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:51.715881+0000 mon.vm03 (mon.0) 379 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:52.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:51.715881+0000 mon.vm03 (mon.0) 379 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:52.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:51.715962+0000 mon.vm03 (mon.0) 380 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:52.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:51.715962+0000 mon.vm03 (mon.0) 380 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:52.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: cluster 2026-03-06T21:17:52.057636+0000 mgr.vm03.uwuzgl (mgr.14199) 70 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:52.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: cluster 2026-03-06T21:17:52.057636+0000 mgr.vm03.uwuzgl (mgr.14199) 70 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:52.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:52.153837+0000 mon.vm08 (mon.1) 6 : audit [DBG] from='client.? 192.168.123.108:0/2772924257' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:52.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:52.153837+0000 mon.vm08 (mon.1) 6 : audit [DBG] from='client.? 192.168.123.108:0/2772924257' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:52.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:52.339286+0000 mon.vm03 (mon.0) 381 : audit [DBG] from='client.? 192.168.123.103:0/3702303720' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:52.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:52.339286+0000 mon.vm03 (mon.0) 381 : audit [DBG] from='client.? 192.168.123.103:0/3702303720' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:52.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:52.407035+0000 mon.vm03 (mon.0) 382 : audit [DBG] from='client.? 192.168.123.103:0/3526956232' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:17:52.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:52 vm03 bash[17055]: audit 2026-03-06T21:17:52.407035+0000 mon.vm03 (mon.0) 382 : audit [DBG] from='client.? 192.168.123.103:0/3526956232' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:51.710265+0000 mon.vm03 (mon.0) 374 : audit [INF] from='client.? 192.168.123.103:0/2076291687' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8bb1dfca-f66a-4fa1-86a3-bb356250604f"}]: dispatch 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:51.710265+0000 mon.vm03 (mon.0) 374 : audit [INF] from='client.? 192.168.123.103:0/2076291687' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8bb1dfca-f66a-4fa1-86a3-bb356250604f"}]: dispatch 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:51.713271+0000 mon.vm03 (mon.0) 375 : audit [INF] from='client.? 192.168.123.103:0/2076291687' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8bb1dfca-f66a-4fa1-86a3-bb356250604f"}]': finished 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:51.713271+0000 mon.vm03 (mon.0) 375 : audit [INF] from='client.? 192.168.123.103:0/2076291687' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8bb1dfca-f66a-4fa1-86a3-bb356250604f"}]': finished 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: cluster 2026-03-06T21:17:51.715557+0000 mon.vm03 (mon.0) 376 : cluster [DBG] osdmap e9: 4 total, 0 up, 4 in 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: cluster 2026-03-06T21:17:51.715557+0000 mon.vm03 (mon.0) 376 : cluster [DBG] osdmap e9: 4 total, 0 up, 4 in 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:51.715708+0000 mon.vm03 (mon.0) 377 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:51.715708+0000 mon.vm03 (mon.0) 377 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:51.715801+0000 mon.vm03 (mon.0) 378 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:51.715801+0000 mon.vm03 (mon.0) 378 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:51.715881+0000 mon.vm03 (mon.0) 379 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:51.715881+0000 mon.vm03 (mon.0) 379 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:51.715962+0000 mon.vm03 (mon.0) 380 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:51.715962+0000 mon.vm03 (mon.0) 380 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: cluster 2026-03-06T21:17:52.057636+0000 mgr.vm03.uwuzgl (mgr.14199) 70 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: cluster 2026-03-06T21:17:52.057636+0000 mgr.vm03.uwuzgl (mgr.14199) 70 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:52.153837+0000 mon.vm08 (mon.1) 6 : audit [DBG] from='client.? 192.168.123.108:0/2772924257' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:52.153837+0000 mon.vm08 (mon.1) 6 : audit [DBG] from='client.? 192.168.123.108:0/2772924257' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:52.339286+0000 mon.vm03 (mon.0) 381 : audit [DBG] from='client.? 192.168.123.103:0/3702303720' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:53.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:52.339286+0000 mon.vm03 (mon.0) 381 : audit [DBG] from='client.? 192.168.123.103:0/3702303720' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:53.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:52.407035+0000 mon.vm03 (mon.0) 382 : audit [DBG] from='client.? 192.168.123.103:0/3526956232' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:17:53.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:52 vm08 bash[20828]: audit 2026-03-06T21:17:52.407035+0000 mon.vm03 (mon.0) 382 : audit [DBG] from='client.? 192.168.123.103:0/3526956232' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:17:53.469 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd stat -f json 2026-03-06T22:17:55.373 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:55 vm08 bash[20828]: cluster 2026-03-06T21:17:54.057817+0000 mgr.vm03.uwuzgl (mgr.14199) 71 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:55.373 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:55 vm08 bash[20828]: cluster 2026-03-06T21:17:54.057817+0000 mgr.vm03.uwuzgl (mgr.14199) 71 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:55.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:55 vm03 bash[17055]: cluster 2026-03-06T21:17:54.057817+0000 mgr.vm03.uwuzgl (mgr.14199) 71 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:55.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:55 vm03 bash[17055]: cluster 2026-03-06T21:17:54.057817+0000 mgr.vm03.uwuzgl (mgr.14199) 71 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:56.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.241767+0000 mon.vm03 (mon.0) 383 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d184c72d-8b18-42f5-b09e-8d2a36892d30"}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.241767+0000 mon.vm03 (mon.0) 383 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d184c72d-8b18-42f5-b09e-8d2a36892d30"}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.244213+0000 mon.vm03 (mon.0) 384 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "d184c72d-8b18-42f5-b09e-8d2a36892d30"}]': finished 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.244213+0000 mon.vm03 (mon.0) 384 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "d184c72d-8b18-42f5-b09e-8d2a36892d30"}]': finished 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.245756+0000 mon.vm08 (mon.1) 7 : audit [INF] from='client.? 192.168.123.108:0/3012875908' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d184c72d-8b18-42f5-b09e-8d2a36892d30"}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.245756+0000 mon.vm08 (mon.1) 7 : audit [INF] from='client.? 192.168.123.108:0/3012875908' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d184c72d-8b18-42f5-b09e-8d2a36892d30"}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: cluster 2026-03-06T21:17:55.246024+0000 mon.vm03 (mon.0) 385 : cluster [DBG] osdmap e10: 5 total, 0 up, 5 in 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: cluster 2026-03-06T21:17:55.246024+0000 mon.vm03 (mon.0) 385 : cluster [DBG] osdmap e10: 5 total, 0 up, 5 in 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.246279+0000 mon.vm03 (mon.0) 386 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.246279+0000 mon.vm03 (mon.0) 386 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.246393+0000 mon.vm03 (mon.0) 387 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.246393+0000 mon.vm03 (mon.0) 387 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.246520+0000 mon.vm03 (mon.0) 388 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.246520+0000 mon.vm03 (mon.0) 388 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.246622+0000 mon.vm03 (mon.0) 389 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.246622+0000 mon.vm03 (mon.0) 389 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.246727+0000 mon.vm03 (mon.0) 390 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.246727+0000 mon.vm03 (mon.0) 390 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.475334+0000 mon.vm03 (mon.0) 391 : audit [INF] from='client.? 192.168.123.103:0/2163546933' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "58e80b7b-b647-452b-9f23-bf7ae1e3b6bb"}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.475334+0000 mon.vm03 (mon.0) 391 : audit [INF] from='client.? 192.168.123.103:0/2163546933' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "58e80b7b-b647-452b-9f23-bf7ae1e3b6bb"}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.477932+0000 mon.vm03 (mon.0) 392 : audit [INF] from='client.? 192.168.123.103:0/2163546933' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "58e80b7b-b647-452b-9f23-bf7ae1e3b6bb"}]': finished 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.477932+0000 mon.vm03 (mon.0) 392 : audit [INF] from='client.? 192.168.123.103:0/2163546933' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "58e80b7b-b647-452b-9f23-bf7ae1e3b6bb"}]': finished 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: cluster 2026-03-06T21:17:55.479763+0000 mon.vm03 (mon.0) 393 : cluster [DBG] osdmap e11: 6 total, 0 up, 6 in 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: cluster 2026-03-06T21:17:55.479763+0000 mon.vm03 (mon.0) 393 : cluster [DBG] osdmap e11: 6 total, 0 up, 6 in 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.479882+0000 mon.vm03 (mon.0) 394 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.479882+0000 mon.vm03 (mon.0) 394 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.479953+0000 mon.vm03 (mon.0) 395 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.479953+0000 mon.vm03 (mon.0) 395 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.480076+0000 mon.vm03 (mon.0) 396 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.480076+0000 mon.vm03 (mon.0) 396 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.480190+0000 mon.vm03 (mon.0) 397 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.480190+0000 mon.vm03 (mon.0) 397 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.480291+0000 mon.vm03 (mon.0) 398 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.480291+0000 mon.vm03 (mon.0) 398 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.480391+0000 mon.vm03 (mon.0) 399 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.480391+0000 mon.vm03 (mon.0) 399 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.847221+0000 mon.vm08 (mon.1) 8 : audit [DBG] from='client.? 192.168.123.108:0/1533822316' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:55.847221+0000 mon.vm08 (mon.1) 8 : audit [DBG] from='client.? 192.168.123.108:0/1533822316' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:56.061735+0000 mon.vm03 (mon.0) 400 : audit [DBG] from='client.? 192.168.123.103:0/3060466239' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:56.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:56 vm03 bash[17055]: audit 2026-03-06T21:17:56.061735+0000 mon.vm03 (mon.0) 400 : audit [DBG] from='client.? 192.168.123.103:0/3060466239' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:56.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.241767+0000 mon.vm03 (mon.0) 383 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d184c72d-8b18-42f5-b09e-8d2a36892d30"}]: dispatch 2026-03-06T22:17:56.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.241767+0000 mon.vm03 (mon.0) 383 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d184c72d-8b18-42f5-b09e-8d2a36892d30"}]: dispatch 2026-03-06T22:17:56.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.244213+0000 mon.vm03 (mon.0) 384 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "d184c72d-8b18-42f5-b09e-8d2a36892d30"}]': finished 2026-03-06T22:17:56.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.244213+0000 mon.vm03 (mon.0) 384 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "d184c72d-8b18-42f5-b09e-8d2a36892d30"}]': finished 2026-03-06T22:17:56.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.245756+0000 mon.vm08 (mon.1) 7 : audit [INF] from='client.? 192.168.123.108:0/3012875908' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d184c72d-8b18-42f5-b09e-8d2a36892d30"}]: dispatch 2026-03-06T22:17:56.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.245756+0000 mon.vm08 (mon.1) 7 : audit [INF] from='client.? 192.168.123.108:0/3012875908' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d184c72d-8b18-42f5-b09e-8d2a36892d30"}]: dispatch 2026-03-06T22:17:56.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: cluster 2026-03-06T21:17:55.246024+0000 mon.vm03 (mon.0) 385 : cluster [DBG] osdmap e10: 5 total, 0 up, 5 in 2026-03-06T22:17:56.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: cluster 2026-03-06T21:17:55.246024+0000 mon.vm03 (mon.0) 385 : cluster [DBG] osdmap e10: 5 total, 0 up, 5 in 2026-03-06T22:17:56.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.246279+0000 mon.vm03 (mon.0) 386 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:56.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.246279+0000 mon.vm03 (mon.0) 386 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:56.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.246393+0000 mon.vm03 (mon.0) 387 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:56.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.246393+0000 mon.vm03 (mon.0) 387 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:56.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.246520+0000 mon.vm03 (mon.0) 388 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.246520+0000 mon.vm03 (mon.0) 388 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.246622+0000 mon.vm03 (mon.0) 389 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.246622+0000 mon.vm03 (mon.0) 389 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.246727+0000 mon.vm03 (mon.0) 390 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.246727+0000 mon.vm03 (mon.0) 390 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.475334+0000 mon.vm03 (mon.0) 391 : audit [INF] from='client.? 192.168.123.103:0/2163546933' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "58e80b7b-b647-452b-9f23-bf7ae1e3b6bb"}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.475334+0000 mon.vm03 (mon.0) 391 : audit [INF] from='client.? 192.168.123.103:0/2163546933' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "58e80b7b-b647-452b-9f23-bf7ae1e3b6bb"}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.477932+0000 mon.vm03 (mon.0) 392 : audit [INF] from='client.? 192.168.123.103:0/2163546933' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "58e80b7b-b647-452b-9f23-bf7ae1e3b6bb"}]': finished 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.477932+0000 mon.vm03 (mon.0) 392 : audit [INF] from='client.? 192.168.123.103:0/2163546933' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "58e80b7b-b647-452b-9f23-bf7ae1e3b6bb"}]': finished 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: cluster 2026-03-06T21:17:55.479763+0000 mon.vm03 (mon.0) 393 : cluster [DBG] osdmap e11: 6 total, 0 up, 6 in 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: cluster 2026-03-06T21:17:55.479763+0000 mon.vm03 (mon.0) 393 : cluster [DBG] osdmap e11: 6 total, 0 up, 6 in 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.479882+0000 mon.vm03 (mon.0) 394 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.479882+0000 mon.vm03 (mon.0) 394 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.479953+0000 mon.vm03 (mon.0) 395 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.479953+0000 mon.vm03 (mon.0) 395 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.480076+0000 mon.vm03 (mon.0) 396 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.480076+0000 mon.vm03 (mon.0) 396 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.480190+0000 mon.vm03 (mon.0) 397 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.480190+0000 mon.vm03 (mon.0) 397 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.480291+0000 mon.vm03 (mon.0) 398 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.480291+0000 mon.vm03 (mon.0) 398 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.480391+0000 mon.vm03 (mon.0) 399 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.480391+0000 mon.vm03 (mon.0) 399 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.847221+0000 mon.vm08 (mon.1) 8 : audit [DBG] from='client.? 192.168.123.108:0/1533822316' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:55.847221+0000 mon.vm08 (mon.1) 8 : audit [DBG] from='client.? 192.168.123.108:0/1533822316' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:56.061735+0000 mon.vm03 (mon.0) 400 : audit [DBG] from='client.? 192.168.123.103:0/3060466239' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:56.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:56 vm08 bash[20828]: audit 2026-03-06T21:17:56.061735+0000 mon.vm03 (mon.0) 400 : audit [DBG] from='client.? 192.168.123.103:0/3060466239' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:17:57.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:57 vm08 bash[20828]: cluster 2026-03-06T21:17:56.057985+0000 mgr.vm03.uwuzgl (mgr.14199) 72 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:57.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:57 vm08 bash[20828]: cluster 2026-03-06T21:17:56.057985+0000 mgr.vm03.uwuzgl (mgr.14199) 72 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:57.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:57 vm03 bash[17055]: cluster 2026-03-06T21:17:56.057985+0000 mgr.vm03.uwuzgl (mgr.14199) 72 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:57.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:57 vm03 bash[17055]: cluster 2026-03-06T21:17:56.057985+0000 mgr.vm03.uwuzgl (mgr.14199) 72 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:58.235 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:17:58.582 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:17:58.638 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":11,"num_osds":6,"num_up_osds":0,"osd_up_since":0,"num_in_osds":6,"osd_in_since":1772831875,"num_remapped_pgs":0} 2026-03-06T22:17:59.639 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd stat -f json 2026-03-06T22:17:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: cluster 2026-03-06T21:17:58.058473+0000 mgr.vm03.uwuzgl (mgr.14199) 73 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: cluster 2026-03-06T21:17:58.058473+0000 mgr.vm03.uwuzgl (mgr.14199) 73 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.576950+0000 mon.vm03 (mon.0) 401 : audit [DBG] from='client.? 192.168.123.103:0/1306028397' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:17:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.576950+0000 mon.vm03 (mon.0) 401 : audit [DBG] from='client.? 192.168.123.103:0/1306028397' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:17:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.939617+0000 mon.vm03 (mon.0) 402 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "bc2367fc-eb8e-4dec-9239-555bf4ed1271"}]: dispatch 2026-03-06T22:17:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.939617+0000 mon.vm03 (mon.0) 402 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "bc2367fc-eb8e-4dec-9239-555bf4ed1271"}]: dispatch 2026-03-06T22:17:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.942426+0000 mon.vm03 (mon.0) 403 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "bc2367fc-eb8e-4dec-9239-555bf4ed1271"}]': finished 2026-03-06T22:17:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.942426+0000 mon.vm03 (mon.0) 403 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "bc2367fc-eb8e-4dec-9239-555bf4ed1271"}]': finished 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.943786+0000 mon.vm08 (mon.1) 9 : audit [INF] from='client.? 192.168.123.108:0/344669584' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "bc2367fc-eb8e-4dec-9239-555bf4ed1271"}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.943786+0000 mon.vm08 (mon.1) 9 : audit [INF] from='client.? 192.168.123.108:0/344669584' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "bc2367fc-eb8e-4dec-9239-555bf4ed1271"}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: cluster 2026-03-06T21:17:58.944392+0000 mon.vm03 (mon.0) 404 : cluster [DBG] osdmap e12: 7 total, 0 up, 7 in 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: cluster 2026-03-06T21:17:58.944392+0000 mon.vm03 (mon.0) 404 : cluster [DBG] osdmap e12: 7 total, 0 up, 7 in 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.944784+0000 mon.vm03 (mon.0) 405 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.944784+0000 mon.vm03 (mon.0) 405 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.944890+0000 mon.vm03 (mon.0) 406 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.944890+0000 mon.vm03 (mon.0) 406 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.944976+0000 mon.vm03 (mon.0) 407 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.944976+0000 mon.vm03 (mon.0) 407 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.945053+0000 mon.vm03 (mon.0) 408 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.945053+0000 mon.vm03 (mon.0) 408 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.945191+0000 mon.vm03 (mon.0) 409 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.945191+0000 mon.vm03 (mon.0) 409 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.945277+0000 mon.vm03 (mon.0) 410 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.945277+0000 mon.vm03 (mon.0) 410 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.945350+0000 mon.vm03 (mon.0) 411 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:58.945350+0000 mon.vm03 (mon.0) 411 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.429797+0000 mon.vm03 (mon.0) 412 : audit [INF] from='client.? 192.168.123.103:0/3247170245' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4fa45945-b055-4d3b-b99b-47e0ac3cbe06"}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.429797+0000 mon.vm03 (mon.0) 412 : audit [INF] from='client.? 192.168.123.103:0/3247170245' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4fa45945-b055-4d3b-b99b-47e0ac3cbe06"}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.432709+0000 mon.vm03 (mon.0) 413 : audit [INF] from='client.? 192.168.123.103:0/3247170245' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "4fa45945-b055-4d3b-b99b-47e0ac3cbe06"}]': finished 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.432709+0000 mon.vm03 (mon.0) 413 : audit [INF] from='client.? 192.168.123.103:0/3247170245' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "4fa45945-b055-4d3b-b99b-47e0ac3cbe06"}]': finished 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: cluster 2026-03-06T21:17:59.434583+0000 mon.vm03 (mon.0) 414 : cluster [DBG] osdmap e13: 8 total, 0 up, 8 in 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: cluster 2026-03-06T21:17:59.434583+0000 mon.vm03 (mon.0) 414 : cluster [DBG] osdmap e13: 8 total, 0 up, 8 in 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.434912+0000 mon.vm03 (mon.0) 415 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.434912+0000 mon.vm03 (mon.0) 415 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.435078+0000 mon.vm03 (mon.0) 416 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.435078+0000 mon.vm03 (mon.0) 416 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.435161+0000 mon.vm03 (mon.0) 417 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.435161+0000 mon.vm03 (mon.0) 417 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.435245+0000 mon.vm03 (mon.0) 418 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.435245+0000 mon.vm03 (mon.0) 418 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.435327+0000 mon.vm03 (mon.0) 419 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.435327+0000 mon.vm03 (mon.0) 419 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.435404+0000 mon.vm03 (mon.0) 420 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.435404+0000 mon.vm03 (mon.0) 420 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.435665+0000 mon.vm03 (mon.0) 421 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.435665+0000 mon.vm03 (mon.0) 421 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.435977+0000 mon.vm03 (mon.0) 422 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:17:59.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:17:59 vm08 bash[20828]: audit 2026-03-06T21:17:59.435977+0000 mon.vm03 (mon.0) 422 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: cluster 2026-03-06T21:17:58.058473+0000 mgr.vm03.uwuzgl (mgr.14199) 73 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: cluster 2026-03-06T21:17:58.058473+0000 mgr.vm03.uwuzgl (mgr.14199) 73 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.576950+0000 mon.vm03 (mon.0) 401 : audit [DBG] from='client.? 192.168.123.103:0/1306028397' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.576950+0000 mon.vm03 (mon.0) 401 : audit [DBG] from='client.? 192.168.123.103:0/1306028397' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.939617+0000 mon.vm03 (mon.0) 402 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "bc2367fc-eb8e-4dec-9239-555bf4ed1271"}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.939617+0000 mon.vm03 (mon.0) 402 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "bc2367fc-eb8e-4dec-9239-555bf4ed1271"}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.942426+0000 mon.vm03 (mon.0) 403 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "bc2367fc-eb8e-4dec-9239-555bf4ed1271"}]': finished 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.942426+0000 mon.vm03 (mon.0) 403 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "bc2367fc-eb8e-4dec-9239-555bf4ed1271"}]': finished 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.943786+0000 mon.vm08 (mon.1) 9 : audit [INF] from='client.? 192.168.123.108:0/344669584' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "bc2367fc-eb8e-4dec-9239-555bf4ed1271"}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.943786+0000 mon.vm08 (mon.1) 9 : audit [INF] from='client.? 192.168.123.108:0/344669584' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "bc2367fc-eb8e-4dec-9239-555bf4ed1271"}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: cluster 2026-03-06T21:17:58.944392+0000 mon.vm03 (mon.0) 404 : cluster [DBG] osdmap e12: 7 total, 0 up, 7 in 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: cluster 2026-03-06T21:17:58.944392+0000 mon.vm03 (mon.0) 404 : cluster [DBG] osdmap e12: 7 total, 0 up, 7 in 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.944784+0000 mon.vm03 (mon.0) 405 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.944784+0000 mon.vm03 (mon.0) 405 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.944890+0000 mon.vm03 (mon.0) 406 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.944890+0000 mon.vm03 (mon.0) 406 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.944976+0000 mon.vm03 (mon.0) 407 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.944976+0000 mon.vm03 (mon.0) 407 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.945053+0000 mon.vm03 (mon.0) 408 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.945053+0000 mon.vm03 (mon.0) 408 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.945191+0000 mon.vm03 (mon.0) 409 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.945191+0000 mon.vm03 (mon.0) 409 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.945277+0000 mon.vm03 (mon.0) 410 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.945277+0000 mon.vm03 (mon.0) 410 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.945350+0000 mon.vm03 (mon.0) 411 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:58.945350+0000 mon.vm03 (mon.0) 411 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.429797+0000 mon.vm03 (mon.0) 412 : audit [INF] from='client.? 192.168.123.103:0/3247170245' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4fa45945-b055-4d3b-b99b-47e0ac3cbe06"}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.429797+0000 mon.vm03 (mon.0) 412 : audit [INF] from='client.? 192.168.123.103:0/3247170245' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4fa45945-b055-4d3b-b99b-47e0ac3cbe06"}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.432709+0000 mon.vm03 (mon.0) 413 : audit [INF] from='client.? 192.168.123.103:0/3247170245' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "4fa45945-b055-4d3b-b99b-47e0ac3cbe06"}]': finished 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.432709+0000 mon.vm03 (mon.0) 413 : audit [INF] from='client.? 192.168.123.103:0/3247170245' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "4fa45945-b055-4d3b-b99b-47e0ac3cbe06"}]': finished 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: cluster 2026-03-06T21:17:59.434583+0000 mon.vm03 (mon.0) 414 : cluster [DBG] osdmap e13: 8 total, 0 up, 8 in 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: cluster 2026-03-06T21:17:59.434583+0000 mon.vm03 (mon.0) 414 : cluster [DBG] osdmap e13: 8 total, 0 up, 8 in 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.434912+0000 mon.vm03 (mon.0) 415 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.434912+0000 mon.vm03 (mon.0) 415 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.435078+0000 mon.vm03 (mon.0) 416 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.435078+0000 mon.vm03 (mon.0) 416 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.435161+0000 mon.vm03 (mon.0) 417 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.435161+0000 mon.vm03 (mon.0) 417 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.435245+0000 mon.vm03 (mon.0) 418 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.435245+0000 mon.vm03 (mon.0) 418 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:17:59.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.435327+0000 mon.vm03 (mon.0) 419 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:17:59.942 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.435327+0000 mon.vm03 (mon.0) 419 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:17:59.942 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.435404+0000 mon.vm03 (mon.0) 420 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:17:59.942 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.435404+0000 mon.vm03 (mon.0) 420 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:17:59.942 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.435665+0000 mon.vm03 (mon.0) 421 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:17:59.942 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.435665+0000 mon.vm03 (mon.0) 421 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:17:59.942 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.435977+0000 mon.vm03 (mon.0) 422 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:17:59.942 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:17:59 vm03 bash[17055]: audit 2026-03-06T21:17:59.435977+0000 mon.vm03 (mon.0) 422 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:00.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:00 vm08 bash[20828]: audit 2026-03-06T21:17:59.526951+0000 mon.vm08 (mon.1) 10 : audit [DBG] from='client.? 192.168.123.108:0/3641586063' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:18:00.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:00 vm08 bash[20828]: audit 2026-03-06T21:17:59.526951+0000 mon.vm08 (mon.1) 10 : audit [DBG] from='client.? 192.168.123.108:0/3641586063' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:18:00.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:00 vm08 bash[20828]: audit 2026-03-06T21:18:00.055198+0000 mon.vm03 (mon.0) 423 : audit [DBG] from='client.? 192.168.123.103:0/14451677' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:18:00.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:00 vm08 bash[20828]: audit 2026-03-06T21:18:00.055198+0000 mon.vm03 (mon.0) 423 : audit [DBG] from='client.? 192.168.123.103:0/14451677' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:18:00.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:00 vm03 bash[17055]: audit 2026-03-06T21:17:59.526951+0000 mon.vm08 (mon.1) 10 : audit [DBG] from='client.? 192.168.123.108:0/3641586063' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:18:00.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:00 vm03 bash[17055]: audit 2026-03-06T21:17:59.526951+0000 mon.vm08 (mon.1) 10 : audit [DBG] from='client.? 192.168.123.108:0/3641586063' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:18:00.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:00 vm03 bash[17055]: audit 2026-03-06T21:18:00.055198+0000 mon.vm03 (mon.0) 423 : audit [DBG] from='client.? 192.168.123.103:0/14451677' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:18:00.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:00 vm03 bash[17055]: audit 2026-03-06T21:18:00.055198+0000 mon.vm03 (mon.0) 423 : audit [DBG] from='client.? 192.168.123.103:0/14451677' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T22:18:01.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:01 vm03 bash[17055]: cluster 2026-03-06T21:18:00.058643+0000 mgr.vm03.uwuzgl (mgr.14199) 74 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:01.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:01 vm03 bash[17055]: cluster 2026-03-06T21:18:00.058643+0000 mgr.vm03.uwuzgl (mgr.14199) 74 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:02.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:01 vm08 bash[20828]: cluster 2026-03-06T21:18:00.058643+0000 mgr.vm03.uwuzgl (mgr.14199) 74 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:02.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:01 vm08 bash[20828]: cluster 2026-03-06T21:18:00.058643+0000 mgr.vm03.uwuzgl (mgr.14199) 74 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:03.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:03 vm03 bash[17055]: cluster 2026-03-06T21:18:02.058862+0000 mgr.vm03.uwuzgl (mgr.14199) 75 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:03.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:03 vm03 bash[17055]: cluster 2026-03-06T21:18:02.058862+0000 mgr.vm03.uwuzgl (mgr.14199) 75 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:04.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:03 vm08 bash[20828]: cluster 2026-03-06T21:18:02.058862+0000 mgr.vm03.uwuzgl (mgr.14199) 75 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:04.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:03 vm08 bash[20828]: cluster 2026-03-06T21:18:02.058862+0000 mgr.vm03.uwuzgl (mgr.14199) 75 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:04.406 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:18:04.745 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:18:04.814 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1772831879,"num_remapped_pgs":0} 2026-03-06T22:18:05.815 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd stat -f json 2026-03-06T22:18:05.821 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:05 vm03 bash[17055]: cluster 2026-03-06T21:18:04.059064+0000 mgr.vm03.uwuzgl (mgr.14199) 76 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:05.821 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:05 vm03 bash[17055]: cluster 2026-03-06T21:18:04.059064+0000 mgr.vm03.uwuzgl (mgr.14199) 76 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:05.821 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:05 vm03 bash[17055]: audit 2026-03-06T21:18:04.739450+0000 mon.vm03 (mon.0) 424 : audit [DBG] from='client.? 192.168.123.103:0/3271700279' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:18:05.821 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:05 vm03 bash[17055]: audit 2026-03-06T21:18:04.739450+0000 mon.vm03 (mon.0) 424 : audit [DBG] from='client.? 192.168.123.103:0/3271700279' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:18:06.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:05 vm08 bash[20828]: cluster 2026-03-06T21:18:04.059064+0000 mgr.vm03.uwuzgl (mgr.14199) 76 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:06.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:05 vm08 bash[20828]: cluster 2026-03-06T21:18:04.059064+0000 mgr.vm03.uwuzgl (mgr.14199) 76 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:06.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:05 vm08 bash[20828]: audit 2026-03-06T21:18:04.739450+0000 mon.vm03 (mon.0) 424 : audit [DBG] from='client.? 192.168.123.103:0/3271700279' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:18:06.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:05 vm08 bash[20828]: audit 2026-03-06T21:18:04.739450+0000 mon.vm03 (mon.0) 424 : audit [DBG] from='client.? 192.168.123.103:0/3271700279' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:18:06.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:06 vm03 bash[17055]: audit 2026-03-06T21:18:06.056661+0000 mon.vm03 (mon.0) 425 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:18:06.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:06 vm03 bash[17055]: audit 2026-03-06T21:18:06.056661+0000 mon.vm03 (mon.0) 425 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:18:07.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:06 vm08 bash[20828]: audit 2026-03-06T21:18:06.056661+0000 mon.vm03 (mon.0) 425 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:18:07.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:06 vm08 bash[20828]: audit 2026-03-06T21:18:06.056661+0000 mon.vm03 (mon.0) 425 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:18:07.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:07 vm03 bash[17055]: cluster 2026-03-06T21:18:06.066273+0000 mgr.vm03.uwuzgl (mgr.14199) 77 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:07.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:07 vm03 bash[17055]: cluster 2026-03-06T21:18:06.066273+0000 mgr.vm03.uwuzgl (mgr.14199) 77 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:08.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:07 vm08 bash[20828]: cluster 2026-03-06T21:18:06.066273+0000 mgr.vm03.uwuzgl (mgr.14199) 77 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:08.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:07 vm08 bash[20828]: cluster 2026-03-06T21:18:06.066273+0000 mgr.vm03.uwuzgl (mgr.14199) 77 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:08.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:08 vm03 bash[17055]: audit 2026-03-06T21:18:08.104132+0000 mon.vm03 (mon.0) 426 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T22:18:08.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:08 vm03 bash[17055]: audit 2026-03-06T21:18:08.104132+0000 mon.vm03 (mon.0) 426 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T22:18:08.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:08 vm03 bash[17055]: audit 2026-03-06T21:18:08.104667+0000 mon.vm03 (mon.0) 427 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:08.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:08 vm03 bash[17055]: audit 2026-03-06T21:18:08.104667+0000 mon.vm03 (mon.0) 427 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:08.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:08 vm03 bash[17055]: audit 2026-03-06T21:18:08.202481+0000 mon.vm03 (mon.0) 428 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T22:18:08.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:08 vm03 bash[17055]: audit 2026-03-06T21:18:08.202481+0000 mon.vm03 (mon.0) 428 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T22:18:08.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:08 vm03 bash[17055]: audit 2026-03-06T21:18:08.202967+0000 mon.vm03 (mon.0) 429 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:08.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:08 vm03 bash[17055]: audit 2026-03-06T21:18:08.202967+0000 mon.vm03 (mon.0) 429 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:08.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:08 vm08 bash[20828]: audit 2026-03-06T21:18:08.104132+0000 mon.vm03 (mon.0) 426 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T22:18:08.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:08 vm08 bash[20828]: audit 2026-03-06T21:18:08.104132+0000 mon.vm03 (mon.0) 426 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T22:18:08.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:08 vm08 bash[20828]: audit 2026-03-06T21:18:08.104667+0000 mon.vm03 (mon.0) 427 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:08.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:08 vm08 bash[20828]: audit 2026-03-06T21:18:08.104667+0000 mon.vm03 (mon.0) 427 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:08.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:08 vm08 bash[20828]: audit 2026-03-06T21:18:08.202481+0000 mon.vm03 (mon.0) 428 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T22:18:08.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:08 vm08 bash[20828]: audit 2026-03-06T21:18:08.202481+0000 mon.vm03 (mon.0) 428 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T22:18:08.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:08 vm08 bash[20828]: audit 2026-03-06T21:18:08.202967+0000 mon.vm03 (mon.0) 429 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:08.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:08 vm08 bash[20828]: audit 2026-03-06T21:18:08.202967+0000 mon.vm03 (mon.0) 429 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:09.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:09 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:18:09.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:09 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:18:09.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:09 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:18:09.547 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:09 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:18:09.852 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:09 vm03 bash[17055]: cluster 2026-03-06T21:18:08.066472+0000 mgr.vm03.uwuzgl (mgr.14199) 78 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:09.852 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:09 vm03 bash[17055]: cluster 2026-03-06T21:18:08.066472+0000 mgr.vm03.uwuzgl (mgr.14199) 78 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:09.852 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:09 vm03 bash[17055]: cephadm 2026-03-06T21:18:08.105119+0000 mgr.vm03.uwuzgl (mgr.14199) 79 : cephadm [INF] Deploying daemon osd.1 on vm03 2026-03-06T22:18:09.852 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:09 vm03 bash[17055]: cephadm 2026-03-06T21:18:08.105119+0000 mgr.vm03.uwuzgl (mgr.14199) 79 : cephadm [INF] Deploying daemon osd.1 on vm03 2026-03-06T22:18:09.852 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:09 vm03 bash[17055]: cephadm 2026-03-06T21:18:08.203378+0000 mgr.vm03.uwuzgl (mgr.14199) 80 : cephadm [INF] Deploying daemon osd.0 on vm08 2026-03-06T22:18:09.852 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:09 vm03 bash[17055]: cephadm 2026-03-06T21:18:08.203378+0000 mgr.vm03.uwuzgl (mgr.14199) 80 : cephadm [INF] Deploying daemon osd.0 on vm08 2026-03-06T22:18:09.853 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:09 vm08 bash[20828]: cluster 2026-03-06T21:18:08.066472+0000 mgr.vm03.uwuzgl (mgr.14199) 78 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:09.853 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:09 vm08 bash[20828]: cluster 2026-03-06T21:18:08.066472+0000 mgr.vm03.uwuzgl (mgr.14199) 78 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:09.853 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:09 vm08 bash[20828]: cephadm 2026-03-06T21:18:08.105119+0000 mgr.vm03.uwuzgl (mgr.14199) 79 : cephadm [INF] Deploying daemon osd.1 on vm03 2026-03-06T22:18:09.853 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:09 vm08 bash[20828]: cephadm 2026-03-06T21:18:08.105119+0000 mgr.vm03.uwuzgl (mgr.14199) 79 : cephadm [INF] Deploying daemon osd.1 on vm03 2026-03-06T22:18:09.853 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:09 vm08 bash[20828]: cephadm 2026-03-06T21:18:08.203378+0000 mgr.vm03.uwuzgl (mgr.14199) 80 : cephadm [INF] Deploying daemon osd.0 on vm08 2026-03-06T22:18:09.853 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:09 vm08 bash[20828]: cephadm 2026-03-06T21:18:08.203378+0000 mgr.vm03.uwuzgl (mgr.14199) 80 : cephadm [INF] Deploying daemon osd.0 on vm08 2026-03-06T22:18:10.833 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: audit 2026-03-06T21:18:09.566887+0000 mon.vm03 (mon.0) 430 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:10.833 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: audit 2026-03-06T21:18:09.566887+0000 mon.vm03 (mon.0) 430 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:10.833 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: audit 2026-03-06T21:18:09.571371+0000 mon.vm03 (mon.0) 431 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:10.833 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: audit 2026-03-06T21:18:09.571371+0000 mon.vm03 (mon.0) 431 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:10.833 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: audit 2026-03-06T21:18:09.572036+0000 mon.vm03 (mon.0) 432 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-06T22:18:10.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: audit 2026-03-06T21:18:09.572036+0000 mon.vm03 (mon.0) 432 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-06T22:18:10.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: audit 2026-03-06T21:18:09.572480+0000 mon.vm03 (mon.0) 433 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:10.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: audit 2026-03-06T21:18:09.572480+0000 mon.vm03 (mon.0) 433 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:10.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: cephadm 2026-03-06T21:18:09.572800+0000 mgr.vm03.uwuzgl (mgr.14199) 81 : cephadm [INF] Deploying daemon osd.2 on vm08 2026-03-06T22:18:10.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: cephadm 2026-03-06T21:18:09.572800+0000 mgr.vm03.uwuzgl (mgr.14199) 81 : cephadm [INF] Deploying daemon osd.2 on vm08 2026-03-06T22:18:10.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: audit 2026-03-06T21:18:09.578839+0000 mon.vm03 (mon.0) 434 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:10.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: audit 2026-03-06T21:18:09.578839+0000 mon.vm03 (mon.0) 434 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:10.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: audit 2026-03-06T21:18:09.585331+0000 mon.vm03 (mon.0) 435 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:10.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: audit 2026-03-06T21:18:09.585331+0000 mon.vm03 (mon.0) 435 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:10.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: audit 2026-03-06T21:18:09.586140+0000 mon.vm03 (mon.0) 436 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-06T22:18:10.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: audit 2026-03-06T21:18:09.586140+0000 mon.vm03 (mon.0) 436 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-06T22:18:10.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: audit 2026-03-06T21:18:09.586593+0000 mon.vm03 (mon.0) 437 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:10.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: audit 2026-03-06T21:18:09.586593+0000 mon.vm03 (mon.0) 437 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:10.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: cephadm 2026-03-06T21:18:09.586939+0000 mgr.vm03.uwuzgl (mgr.14199) 82 : cephadm [INF] Deploying daemon osd.3 on vm03 2026-03-06T22:18:10.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: cephadm 2026-03-06T21:18:09.586939+0000 mgr.vm03.uwuzgl (mgr.14199) 82 : cephadm [INF] Deploying daemon osd.3 on vm03 2026-03-06T22:18:10.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: cluster 2026-03-06T21:18:10.066658+0000 mgr.vm03.uwuzgl (mgr.14199) 83 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:10.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 bash[20828]: cluster 2026-03-06T21:18:10.066658+0000 mgr.vm03.uwuzgl (mgr.14199) 83 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: audit 2026-03-06T21:18:09.566887+0000 mon.vm03 (mon.0) 430 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: audit 2026-03-06T21:18:09.566887+0000 mon.vm03 (mon.0) 430 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: audit 2026-03-06T21:18:09.571371+0000 mon.vm03 (mon.0) 431 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: audit 2026-03-06T21:18:09.571371+0000 mon.vm03 (mon.0) 431 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: audit 2026-03-06T21:18:09.572036+0000 mon.vm03 (mon.0) 432 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: audit 2026-03-06T21:18:09.572036+0000 mon.vm03 (mon.0) 432 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: audit 2026-03-06T21:18:09.572480+0000 mon.vm03 (mon.0) 433 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: audit 2026-03-06T21:18:09.572480+0000 mon.vm03 (mon.0) 433 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: cephadm 2026-03-06T21:18:09.572800+0000 mgr.vm03.uwuzgl (mgr.14199) 81 : cephadm [INF] Deploying daemon osd.2 on vm08 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: cephadm 2026-03-06T21:18:09.572800+0000 mgr.vm03.uwuzgl (mgr.14199) 81 : cephadm [INF] Deploying daemon osd.2 on vm08 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: audit 2026-03-06T21:18:09.578839+0000 mon.vm03 (mon.0) 434 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: audit 2026-03-06T21:18:09.578839+0000 mon.vm03 (mon.0) 434 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: audit 2026-03-06T21:18:09.585331+0000 mon.vm03 (mon.0) 435 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: audit 2026-03-06T21:18:09.585331+0000 mon.vm03 (mon.0) 435 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: audit 2026-03-06T21:18:09.586140+0000 mon.vm03 (mon.0) 436 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: audit 2026-03-06T21:18:09.586140+0000 mon.vm03 (mon.0) 436 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: audit 2026-03-06T21:18:09.586593+0000 mon.vm03 (mon.0) 437 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:10.850 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: audit 2026-03-06T21:18:09.586593+0000 mon.vm03 (mon.0) 437 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:10.851 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: cephadm 2026-03-06T21:18:09.586939+0000 mgr.vm03.uwuzgl (mgr.14199) 82 : cephadm [INF] Deploying daemon osd.3 on vm03 2026-03-06T22:18:10.851 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: cephadm 2026-03-06T21:18:09.586939+0000 mgr.vm03.uwuzgl (mgr.14199) 82 : cephadm [INF] Deploying daemon osd.3 on vm03 2026-03-06T22:18:10.851 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: cluster 2026-03-06T21:18:10.066658+0000 mgr.vm03.uwuzgl (mgr.14199) 83 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:10.851 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:10 vm03 bash[17055]: cluster 2026-03-06T21:18:10.066658+0000 mgr.vm03.uwuzgl (mgr.14199) 83 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:11.103 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:11 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:18:11.209 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:10 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:18:11.379 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:11 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:18:11.468 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:11 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:18:11.546 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:18:12.122 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:18:12.214 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1772831879,"num_remapped_pgs":0} 2026-03-06T22:18:12.483 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:11.353374+0000 mon.vm03 (mon.0) 438 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:12.483 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:11.353374+0000 mon.vm03 (mon.0) 438 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:12.483 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:11.360628+0000 mon.vm03 (mon.0) 439 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:12.483 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:11.360628+0000 mon.vm03 (mon.0) 439 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:12.483 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:11.361491+0000 mon.vm03 (mon.0) 440 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-06T22:18:12.483 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:11.361491+0000 mon.vm03 (mon.0) 440 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-06T22:18:12.483 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:11.362081+0000 mon.vm03 (mon.0) 441 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:12.483 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:11.362081+0000 mon.vm03 (mon.0) 441 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:12.483 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: cephadm 2026-03-06T21:18:11.362621+0000 mgr.vm03.uwuzgl (mgr.14199) 84 : cephadm [INF] Deploying daemon osd.4 on vm08 2026-03-06T22:18:12.483 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: cephadm 2026-03-06T21:18:11.362621+0000 mgr.vm03.uwuzgl (mgr.14199) 84 : cephadm [INF] Deploying daemon osd.4 on vm08 2026-03-06T22:18:12.483 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:11.416284+0000 mon.vm03 (mon.0) 442 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:12.483 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:11.416284+0000 mon.vm03 (mon.0) 442 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:12.483 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:11.429234+0000 mon.vm03 (mon.0) 443 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:12.483 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:11.429234+0000 mon.vm03 (mon.0) 443 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:12.483 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:11.430625+0000 mon.vm03 (mon.0) 444 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-06T22:18:12.483 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:11.430625+0000 mon.vm03 (mon.0) 444 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-06T22:18:12.483 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:11.431216+0000 mon.vm03 (mon.0) 445 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:12.484 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:11.431216+0000 mon.vm03 (mon.0) 445 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:12.484 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: cephadm 2026-03-06T21:18:11.431773+0000 mgr.vm03.uwuzgl (mgr.14199) 85 : cephadm [INF] Deploying daemon osd.5 on vm03 2026-03-06T22:18:12.484 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: cephadm 2026-03-06T21:18:11.431773+0000 mgr.vm03.uwuzgl (mgr.14199) 85 : cephadm [INF] Deploying daemon osd.5 on vm03 2026-03-06T22:18:12.484 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:12.111291+0000 mon.vm03 (mon.0) 446 : audit [DBG] from='client.? 192.168.123.103:0/3390655899' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:18:12.484 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:12 vm03 bash[17055]: audit 2026-03-06T21:18:12.111291+0000 mon.vm03 (mon.0) 446 : audit [DBG] from='client.? 192.168.123.103:0/3390655899' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:18:12.548 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:11.353374+0000 mon.vm03 (mon.0) 438 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:12.548 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:11.353374+0000 mon.vm03 (mon.0) 438 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:12.548 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:11.360628+0000 mon.vm03 (mon.0) 439 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:12.548 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:11.360628+0000 mon.vm03 (mon.0) 439 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:12.548 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:11.361491+0000 mon.vm03 (mon.0) 440 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-06T22:18:12.548 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:11.361491+0000 mon.vm03 (mon.0) 440 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-06T22:18:12.548 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:11.362081+0000 mon.vm03 (mon.0) 441 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:12.548 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:11.362081+0000 mon.vm03 (mon.0) 441 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:12.548 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: cephadm 2026-03-06T21:18:11.362621+0000 mgr.vm03.uwuzgl (mgr.14199) 84 : cephadm [INF] Deploying daemon osd.4 on vm08 2026-03-06T22:18:12.548 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: cephadm 2026-03-06T21:18:11.362621+0000 mgr.vm03.uwuzgl (mgr.14199) 84 : cephadm [INF] Deploying daemon osd.4 on vm08 2026-03-06T22:18:12.548 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:11.416284+0000 mon.vm03 (mon.0) 442 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:12.548 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:11.416284+0000 mon.vm03 (mon.0) 442 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:12.548 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:11.429234+0000 mon.vm03 (mon.0) 443 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:12.548 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:11.429234+0000 mon.vm03 (mon.0) 443 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:12.548 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:11.430625+0000 mon.vm03 (mon.0) 444 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-06T22:18:12.548 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:11.430625+0000 mon.vm03 (mon.0) 444 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-06T22:18:12.549 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:11.431216+0000 mon.vm03 (mon.0) 445 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:12.549 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:11.431216+0000 mon.vm03 (mon.0) 445 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:12.549 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: cephadm 2026-03-06T21:18:11.431773+0000 mgr.vm03.uwuzgl (mgr.14199) 85 : cephadm [INF] Deploying daemon osd.5 on vm03 2026-03-06T22:18:12.549 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: cephadm 2026-03-06T21:18:11.431773+0000 mgr.vm03.uwuzgl (mgr.14199) 85 : cephadm [INF] Deploying daemon osd.5 on vm03 2026-03-06T22:18:12.549 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:12.111291+0000 mon.vm03 (mon.0) 446 : audit [DBG] from='client.? 192.168.123.103:0/3390655899' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:18:12.549 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 bash[20828]: audit 2026-03-06T21:18:12.111291+0000 mon.vm03 (mon.0) 446 : audit [DBG] from='client.? 192.168.123.103:0/3390655899' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:18:13.081 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:12 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:18:13.216 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd stat -f json 2026-03-06T22:18:13.352 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:13 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:18:13.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:13 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:18:13.479 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:13 vm03 bash[17055]: cluster 2026-03-06T21:18:12.066850+0000 mgr.vm03.uwuzgl (mgr.14199) 86 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:13.479 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:13 vm03 bash[17055]: cluster 2026-03-06T21:18:12.066850+0000 mgr.vm03.uwuzgl (mgr.14199) 86 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:13.479 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:13 vm03 bash[17055]: audit 2026-03-06T21:18:13.209773+0000 mon.vm03 (mon.0) 447 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:13.479 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:13 vm03 bash[17055]: audit 2026-03-06T21:18:13.209773+0000 mon.vm03 (mon.0) 447 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:13.479 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:13 vm03 bash[17055]: audit 2026-03-06T21:18:13.224181+0000 mon.vm03 (mon.0) 448 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:13.479 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:13 vm03 bash[17055]: audit 2026-03-06T21:18:13.224181+0000 mon.vm03 (mon.0) 448 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:13.479 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:13 vm03 bash[17055]: audit 2026-03-06T21:18:13.228686+0000 mon.vm03 (mon.0) 449 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-06T22:18:13.479 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:13 vm03 bash[17055]: audit 2026-03-06T21:18:13.228686+0000 mon.vm03 (mon.0) 449 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-06T22:18:13.479 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:13 vm03 bash[17055]: audit 2026-03-06T21:18:13.229339+0000 mon.vm03 (mon.0) 450 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:13.479 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:13 vm03 bash[17055]: audit 2026-03-06T21:18:13.229339+0000 mon.vm03 (mon.0) 450 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:13.479 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:13 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:18:13.641 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:13 vm08 bash[20828]: cluster 2026-03-06T21:18:12.066850+0000 mgr.vm03.uwuzgl (mgr.14199) 86 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:13.641 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:13 vm08 bash[20828]: cluster 2026-03-06T21:18:12.066850+0000 mgr.vm03.uwuzgl (mgr.14199) 86 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:13.641 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:13 vm08 bash[20828]: audit 2026-03-06T21:18:13.209773+0000 mon.vm03 (mon.0) 447 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:13.641 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:13 vm08 bash[20828]: audit 2026-03-06T21:18:13.209773+0000 mon.vm03 (mon.0) 447 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:13.641 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:13 vm08 bash[20828]: audit 2026-03-06T21:18:13.224181+0000 mon.vm03 (mon.0) 448 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:13.641 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:13 vm08 bash[20828]: audit 2026-03-06T21:18:13.224181+0000 mon.vm03 (mon.0) 448 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:13.641 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:13 vm08 bash[20828]: audit 2026-03-06T21:18:13.228686+0000 mon.vm03 (mon.0) 449 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-06T22:18:13.641 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:13 vm08 bash[20828]: audit 2026-03-06T21:18:13.228686+0000 mon.vm03 (mon.0) 449 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-06T22:18:13.641 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:13 vm08 bash[20828]: audit 2026-03-06T21:18:13.229339+0000 mon.vm03 (mon.0) 450 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:13.641 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:13 vm08 bash[20828]: audit 2026-03-06T21:18:13.229339+0000 mon.vm03 (mon.0) 450 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:14.626 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:14 vm08 bash[20828]: cephadm 2026-03-06T21:18:13.229834+0000 mgr.vm03.uwuzgl (mgr.14199) 87 : cephadm [INF] Deploying daemon osd.6 on vm08 2026-03-06T22:18:14.626 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:14 vm08 bash[20828]: cephadm 2026-03-06T21:18:13.229834+0000 mgr.vm03.uwuzgl (mgr.14199) 87 : cephadm [INF] Deploying daemon osd.6 on vm08 2026-03-06T22:18:14.626 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:14 vm08 bash[20828]: audit 2026-03-06T21:18:13.651716+0000 mon.vm03 (mon.0) 451 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:14.626 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:14 vm08 bash[20828]: audit 2026-03-06T21:18:13.651716+0000 mon.vm03 (mon.0) 451 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:14.626 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:14 vm08 bash[20828]: audit 2026-03-06T21:18:13.658875+0000 mon.vm03 (mon.0) 452 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:14.626 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:14 vm08 bash[20828]: audit 2026-03-06T21:18:13.658875+0000 mon.vm03 (mon.0) 452 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:14.626 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:14 vm08 bash[20828]: audit 2026-03-06T21:18:13.662233+0000 mon.vm03 (mon.0) 453 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-06T22:18:14.626 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:14 vm08 bash[20828]: audit 2026-03-06T21:18:13.662233+0000 mon.vm03 (mon.0) 453 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-06T22:18:14.626 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:14 vm08 bash[20828]: audit 2026-03-06T21:18:13.663201+0000 mon.vm03 (mon.0) 454 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:14.626 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:14 vm08 bash[20828]: audit 2026-03-06T21:18:13.663201+0000 mon.vm03 (mon.0) 454 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:14.648 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:14 vm03 bash[17055]: cephadm 2026-03-06T21:18:13.229834+0000 mgr.vm03.uwuzgl (mgr.14199) 87 : cephadm [INF] Deploying daemon osd.6 on vm08 2026-03-06T22:18:14.648 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:14 vm03 bash[17055]: cephadm 2026-03-06T21:18:13.229834+0000 mgr.vm03.uwuzgl (mgr.14199) 87 : cephadm [INF] Deploying daemon osd.6 on vm08 2026-03-06T22:18:14.648 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:14 vm03 bash[17055]: audit 2026-03-06T21:18:13.651716+0000 mon.vm03 (mon.0) 451 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:14.648 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:14 vm03 bash[17055]: audit 2026-03-06T21:18:13.651716+0000 mon.vm03 (mon.0) 451 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:14.648 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:14 vm03 bash[17055]: audit 2026-03-06T21:18:13.658875+0000 mon.vm03 (mon.0) 452 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:14.648 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:14 vm03 bash[17055]: audit 2026-03-06T21:18:13.658875+0000 mon.vm03 (mon.0) 452 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:14.648 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:14 vm03 bash[17055]: audit 2026-03-06T21:18:13.662233+0000 mon.vm03 (mon.0) 453 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-06T22:18:14.648 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:14 vm03 bash[17055]: audit 2026-03-06T21:18:13.662233+0000 mon.vm03 (mon.0) 453 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-06T22:18:14.648 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:14 vm03 bash[17055]: audit 2026-03-06T21:18:13.663201+0000 mon.vm03 (mon.0) 454 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:14.648 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:14 vm03 bash[17055]: audit 2026-03-06T21:18:13.663201+0000 mon.vm03 (mon.0) 454 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:14.933 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:14 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:18:15.193 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:15 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:18:15.574 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:15 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:18:15.574 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:15 vm03 bash[17055]: cephadm 2026-03-06T21:18:13.663724+0000 mgr.vm03.uwuzgl (mgr.14199) 88 : cephadm [INF] Deploying daemon osd.7 on vm03 2026-03-06T22:18:15.574 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:15 vm03 bash[17055]: cephadm 2026-03-06T21:18:13.663724+0000 mgr.vm03.uwuzgl (mgr.14199) 88 : cephadm [INF] Deploying daemon osd.7 on vm03 2026-03-06T22:18:15.574 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:15 vm03 bash[17055]: cluster 2026-03-06T21:18:14.068369+0000 mgr.vm03.uwuzgl (mgr.14199) 89 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:15.574 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:15 vm03 bash[17055]: cluster 2026-03-06T21:18:14.068369+0000 mgr.vm03.uwuzgl (mgr.14199) 89 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:15.574 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:15 vm03 bash[17055]: audit 2026-03-06T21:18:14.654415+0000 mon.vm03 (mon.0) 455 : audit [INF] from='osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T22:18:15.574 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:15 vm03 bash[17055]: audit 2026-03-06T21:18:14.654415+0000 mon.vm03 (mon.0) 455 : audit [INF] from='osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T22:18:15.574 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:15 vm03 bash[17055]: audit 2026-03-06T21:18:14.731579+0000 mon.vm03 (mon.0) 456 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T22:18:15.574 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:15 vm03 bash[17055]: audit 2026-03-06T21:18:14.731579+0000 mon.vm03 (mon.0) 456 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T22:18:15.574 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:15 vm03 bash[17055]: audit 2026-03-06T21:18:14.735560+0000 mon.vm08 (mon.1) 11 : audit [INF] from='osd.0 [v2:192.168.123.108:6800/956656304,v1:192.168.123.108:6801/956656304]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T22:18:15.574 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:15 vm03 bash[17055]: audit 2026-03-06T21:18:14.735560+0000 mon.vm08 (mon.1) 11 : audit [INF] from='osd.0 [v2:192.168.123.108:6800/956656304,v1:192.168.123.108:6801/956656304]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T22:18:15.574 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:15 vm03 bash[17055]: audit 2026-03-06T21:18:15.244570+0000 mon.vm03 (mon.0) 457 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:15.574 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:15 vm03 bash[17055]: audit 2026-03-06T21:18:15.244570+0000 mon.vm03 (mon.0) 457 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:15.574 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:15 vm03 bash[17055]: audit 2026-03-06T21:18:15.253238+0000 mon.vm03 (mon.0) 458 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:15.574 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:15 vm03 bash[17055]: audit 2026-03-06T21:18:15.253238+0000 mon.vm03 (mon.0) 458 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:15.911 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:15 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:18:15.940 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:15 vm08 bash[20828]: cephadm 2026-03-06T21:18:13.663724+0000 mgr.vm03.uwuzgl (mgr.14199) 88 : cephadm [INF] Deploying daemon osd.7 on vm03 2026-03-06T22:18:15.940 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:15 vm08 bash[20828]: cephadm 2026-03-06T21:18:13.663724+0000 mgr.vm03.uwuzgl (mgr.14199) 88 : cephadm [INF] Deploying daemon osd.7 on vm03 2026-03-06T22:18:15.940 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:15 vm08 bash[20828]: cluster 2026-03-06T21:18:14.068369+0000 mgr.vm03.uwuzgl (mgr.14199) 89 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:15.940 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:15 vm08 bash[20828]: cluster 2026-03-06T21:18:14.068369+0000 mgr.vm03.uwuzgl (mgr.14199) 89 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:15.940 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:15 vm08 bash[20828]: audit 2026-03-06T21:18:14.654415+0000 mon.vm03 (mon.0) 455 : audit [INF] from='osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T22:18:15.940 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:15 vm08 bash[20828]: audit 2026-03-06T21:18:14.654415+0000 mon.vm03 (mon.0) 455 : audit [INF] from='osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T22:18:15.940 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:15 vm08 bash[20828]: audit 2026-03-06T21:18:14.731579+0000 mon.vm03 (mon.0) 456 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T22:18:15.940 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:15 vm08 bash[20828]: audit 2026-03-06T21:18:14.731579+0000 mon.vm03 (mon.0) 456 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T22:18:15.940 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:15 vm08 bash[20828]: audit 2026-03-06T21:18:14.735560+0000 mon.vm08 (mon.1) 11 : audit [INF] from='osd.0 [v2:192.168.123.108:6800/956656304,v1:192.168.123.108:6801/956656304]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T22:18:15.940 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:15 vm08 bash[20828]: audit 2026-03-06T21:18:14.735560+0000 mon.vm08 (mon.1) 11 : audit [INF] from='osd.0 [v2:192.168.123.108:6800/956656304,v1:192.168.123.108:6801/956656304]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T22:18:15.940 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:15 vm08 bash[20828]: audit 2026-03-06T21:18:15.244570+0000 mon.vm03 (mon.0) 457 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:15.940 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:15 vm08 bash[20828]: audit 2026-03-06T21:18:15.244570+0000 mon.vm03 (mon.0) 457 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:15.941 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:15 vm08 bash[20828]: audit 2026-03-06T21:18:15.253238+0000 mon.vm03 (mon.0) 458 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:15.941 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:15 vm08 bash[20828]: audit 2026-03-06T21:18:15.253238+0000 mon.vm03 (mon.0) 458 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.459006+0000 mon.vm03 (mon.0) 459 : audit [INF] from='osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.459006+0000 mon.vm03 (mon.0) 459 : audit [INF] from='osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.459109+0000 mon.vm03 (mon.0) 460 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.459109+0000 mon.vm03 (mon.0) 460 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: cluster 2026-03-06T21:18:15.461711+0000 mon.vm03 (mon.0) 461 : cluster [DBG] osdmap e14: 8 total, 0 up, 8 in 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: cluster 2026-03-06T21:18:15.461711+0000 mon.vm03 (mon.0) 461 : cluster [DBG] osdmap e14: 8 total, 0 up, 8 in 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462433+0000 mon.vm03 (mon.0) 462 : audit [INF] from='osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462433+0000 mon.vm03 (mon.0) 462 : audit [INF] from='osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462522+0000 mon.vm03 (mon.0) 463 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462522+0000 mon.vm03 (mon.0) 463 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462567+0000 mon.vm03 (mon.0) 464 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462567+0000 mon.vm03 (mon.0) 464 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462596+0000 mon.vm03 (mon.0) 465 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462596+0000 mon.vm03 (mon.0) 465 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462626+0000 mon.vm03 (mon.0) 466 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462626+0000 mon.vm03 (mon.0) 466 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462653+0000 mon.vm03 (mon.0) 467 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462653+0000 mon.vm03 (mon.0) 467 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462915+0000 mon.vm03 (mon.0) 468 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462915+0000 mon.vm03 (mon.0) 468 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462951+0000 mon.vm03 (mon.0) 469 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462951+0000 mon.vm03 (mon.0) 469 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462982+0000 mon.vm03 (mon.0) 470 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.462982+0000 mon.vm03 (mon.0) 470 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:16.694 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.484760+0000 mon.vm08 (mon.1) 12 : audit [INF] from='osd.0 [v2:192.168.123.108:6800/956656304,v1:192.168.123.108:6801/956656304]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.484760+0000 mon.vm08 (mon.1) 12 : audit [INF] from='osd.0 [v2:192.168.123.108:6800/956656304,v1:192.168.123.108:6801/956656304]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.524804+0000 mon.vm03 (mon.0) 471 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.524804+0000 mon.vm03 (mon.0) 471 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.711462+0000 mon.vm03 (mon.0) 472 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.711462+0000 mon.vm03 (mon.0) 472 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.716754+0000 mon.vm03 (mon.0) 473 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:15.716754+0000 mon.vm03 (mon.0) 473 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.465497+0000 mon.vm03 (mon.0) 474 : audit [INF] from='osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.465497+0000 mon.vm03 (mon.0) 474 : audit [INF] from='osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.465604+0000 mon.vm03 (mon.0) 475 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.465604+0000 mon.vm03 (mon.0) 475 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: cluster 2026-03-06T21:18:16.468943+0000 mon.vm03 (mon.0) 476 : cluster [DBG] osdmap e15: 8 total, 0 up, 8 in 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: cluster 2026-03-06T21:18:16.468943+0000 mon.vm03 (mon.0) 476 : cluster [DBG] osdmap e15: 8 total, 0 up, 8 in 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.469416+0000 mon.vm03 (mon.0) 477 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.469416+0000 mon.vm03 (mon.0) 477 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.469474+0000 mon.vm03 (mon.0) 478 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.469474+0000 mon.vm03 (mon.0) 478 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.469503+0000 mon.vm03 (mon.0) 479 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.469503+0000 mon.vm03 (mon.0) 479 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.469529+0000 mon.vm03 (mon.0) 480 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.469529+0000 mon.vm03 (mon.0) 480 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.469556+0000 mon.vm03 (mon.0) 481 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.469556+0000 mon.vm03 (mon.0) 481 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.469582+0000 mon.vm03 (mon.0) 482 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.469582+0000 mon.vm03 (mon.0) 482 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.469614+0000 mon.vm03 (mon.0) 483 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.469614+0000 mon.vm03 (mon.0) 483 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.470102+0000 mon.vm03 (mon.0) 484 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.470102+0000 mon.vm03 (mon.0) 484 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.471498+0000 mon.vm03 (mon.0) 485 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.471498+0000 mon.vm03 (mon.0) 485 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.475881+0000 mon.vm03 (mon.0) 486 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:18:16.695 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:16 vm08 bash[20828]: audit 2026-03-06T21:18:16.475881+0000 mon.vm03 (mon.0) 486 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:18:16.767 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.459006+0000 mon.vm03 (mon.0) 459 : audit [INF] from='osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-06T22:18:16.767 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.459006+0000 mon.vm03 (mon.0) 459 : audit [INF] from='osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-06T22:18:16.767 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.459109+0000 mon.vm03 (mon.0) 460 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.459109+0000 mon.vm03 (mon.0) 460 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: cluster 2026-03-06T21:18:15.461711+0000 mon.vm03 (mon.0) 461 : cluster [DBG] osdmap e14: 8 total, 0 up, 8 in 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: cluster 2026-03-06T21:18:15.461711+0000 mon.vm03 (mon.0) 461 : cluster [DBG] osdmap e14: 8 total, 0 up, 8 in 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462433+0000 mon.vm03 (mon.0) 462 : audit [INF] from='osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462433+0000 mon.vm03 (mon.0) 462 : audit [INF] from='osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462522+0000 mon.vm03 (mon.0) 463 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462522+0000 mon.vm03 (mon.0) 463 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462567+0000 mon.vm03 (mon.0) 464 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462567+0000 mon.vm03 (mon.0) 464 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462596+0000 mon.vm03 (mon.0) 465 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462596+0000 mon.vm03 (mon.0) 465 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462626+0000 mon.vm03 (mon.0) 466 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462626+0000 mon.vm03 (mon.0) 466 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462653+0000 mon.vm03 (mon.0) 467 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462653+0000 mon.vm03 (mon.0) 467 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462915+0000 mon.vm03 (mon.0) 468 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462915+0000 mon.vm03 (mon.0) 468 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462951+0000 mon.vm03 (mon.0) 469 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462951+0000 mon.vm03 (mon.0) 469 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462982+0000 mon.vm03 (mon.0) 470 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.462982+0000 mon.vm03 (mon.0) 470 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.484760+0000 mon.vm08 (mon.1) 12 : audit [INF] from='osd.0 [v2:192.168.123.108:6800/956656304,v1:192.168.123.108:6801/956656304]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.484760+0000 mon.vm08 (mon.1) 12 : audit [INF] from='osd.0 [v2:192.168.123.108:6800/956656304,v1:192.168.123.108:6801/956656304]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.524804+0000 mon.vm03 (mon.0) 471 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.524804+0000 mon.vm03 (mon.0) 471 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.711462+0000 mon.vm03 (mon.0) 472 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.711462+0000 mon.vm03 (mon.0) 472 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.716754+0000 mon.vm03 (mon.0) 473 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:15.716754+0000 mon.vm03 (mon.0) 473 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.465497+0000 mon.vm03 (mon.0) 474 : audit [INF] from='osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.465497+0000 mon.vm03 (mon.0) 474 : audit [INF] from='osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.465604+0000 mon.vm03 (mon.0) 475 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.465604+0000 mon.vm03 (mon.0) 475 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: cluster 2026-03-06T21:18:16.468943+0000 mon.vm03 (mon.0) 476 : cluster [DBG] osdmap e15: 8 total, 0 up, 8 in 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: cluster 2026-03-06T21:18:16.468943+0000 mon.vm03 (mon.0) 476 : cluster [DBG] osdmap e15: 8 total, 0 up, 8 in 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.469416+0000 mon.vm03 (mon.0) 477 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.469416+0000 mon.vm03 (mon.0) 477 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.469474+0000 mon.vm03 (mon.0) 478 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.469474+0000 mon.vm03 (mon.0) 478 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.469503+0000 mon.vm03 (mon.0) 479 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.469503+0000 mon.vm03 (mon.0) 479 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.469529+0000 mon.vm03 (mon.0) 480 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.469529+0000 mon.vm03 (mon.0) 480 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.469556+0000 mon.vm03 (mon.0) 481 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.469556+0000 mon.vm03 (mon.0) 481 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.469582+0000 mon.vm03 (mon.0) 482 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.469582+0000 mon.vm03 (mon.0) 482 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.469614+0000 mon.vm03 (mon.0) 483 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.469614+0000 mon.vm03 (mon.0) 483 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.470102+0000 mon.vm03 (mon.0) 484 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.470102+0000 mon.vm03 (mon.0) 484 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.471498+0000 mon.vm03 (mon.0) 485 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.471498+0000 mon.vm03 (mon.0) 485 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.475881+0000 mon.vm03 (mon.0) 486 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:18:16.768 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:16 vm03 bash[17055]: audit 2026-03-06T21:18:16.475881+0000 mon.vm03 (mon.0) 486 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:18:17.737 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: cluster 2026-03-06T21:18:16.068572+0000 mgr.vm03.uwuzgl (mgr.14199) 90 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:17.737 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: cluster 2026-03-06T21:18:16.068572+0000 mgr.vm03.uwuzgl (mgr.14199) 90 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:17.737 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:16.879733+0000 mon.vm03 (mon.0) 487 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:16.879733+0000 mon.vm03 (mon.0) 487 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:16.883535+0000 mon.vm08 (mon.1) 13 : audit [INF] from='osd.2 [v2:192.168.123.108:6808/887940165,v1:192.168.123.108:6809/887940165]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:16.883535+0000 mon.vm08 (mon.1) 13 : audit [INF] from='osd.2 [v2:192.168.123.108:6808/887940165,v1:192.168.123.108:6809/887940165]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.468783+0000 mon.vm03 (mon.0) 488 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.468783+0000 mon.vm03 (mon.0) 488 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: cluster 2026-03-06T21:18:17.473232+0000 mon.vm03 (mon.0) 489 : cluster [INF] osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367] boot 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: cluster 2026-03-06T21:18:17.473232+0000 mon.vm03 (mon.0) 489 : cluster [INF] osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367] boot 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: cluster 2026-03-06T21:18:17.473341+0000 mon.vm03 (mon.0) 490 : cluster [INF] osd.0 [v2:192.168.123.108:6800/956656304,v1:192.168.123.108:6801/956656304] boot 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: cluster 2026-03-06T21:18:17.473341+0000 mon.vm03 (mon.0) 490 : cluster [INF] osd.0 [v2:192.168.123.108:6800/956656304,v1:192.168.123.108:6801/956656304] boot 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: cluster 2026-03-06T21:18:17.473445+0000 mon.vm03 (mon.0) 491 : cluster [DBG] osdmap e16: 8 total, 2 up, 8 in 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: cluster 2026-03-06T21:18:17.473445+0000 mon.vm03 (mon.0) 491 : cluster [DBG] osdmap e16: 8 total, 2 up, 8 in 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.473828+0000 mon.vm03 (mon.0) 492 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.473828+0000 mon.vm03 (mon.0) 492 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.473969+0000 mon.vm03 (mon.0) 493 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.473969+0000 mon.vm03 (mon.0) 493 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.474243+0000 mon.vm03 (mon.0) 494 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.474243+0000 mon.vm03 (mon.0) 494 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.474379+0000 mon.vm03 (mon.0) 495 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.474379+0000 mon.vm03 (mon.0) 495 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.474745+0000 mon.vm03 (mon.0) 496 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.474745+0000 mon.vm03 (mon.0) 496 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.474851+0000 mon.vm03 (mon.0) 497 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.474851+0000 mon.vm03 (mon.0) 497 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.474898+0000 mon.vm03 (mon.0) 498 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.474898+0000 mon.vm03 (mon.0) 498 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.474939+0000 mon.vm03 (mon.0) 499 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.474939+0000 mon.vm03 (mon.0) 499 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.475137+0000 mon.vm03 (mon.0) 500 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.475137+0000 mon.vm03 (mon.0) 500 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.477189+0000 mon.vm08 (mon.1) 14 : audit [INF] from='osd.2 [v2:192.168.123.108:6808/887940165,v1:192.168.123.108:6809/887940165]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:17.738 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:17 vm08 bash[20828]: audit 2026-03-06T21:18:17.477189+0000 mon.vm08 (mon.1) 14 : audit [INF] from='osd.2 [v2:192.168.123.108:6808/887940165,v1:192.168.123.108:6809/887940165]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: cluster 2026-03-06T21:18:16.068572+0000 mgr.vm03.uwuzgl (mgr.14199) 90 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: cluster 2026-03-06T21:18:16.068572+0000 mgr.vm03.uwuzgl (mgr.14199) 90 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:16.879733+0000 mon.vm03 (mon.0) 487 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:16.879733+0000 mon.vm03 (mon.0) 487 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:16.883535+0000 mon.vm08 (mon.1) 13 : audit [INF] from='osd.2 [v2:192.168.123.108:6808/887940165,v1:192.168.123.108:6809/887940165]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:16.883535+0000 mon.vm08 (mon.1) 13 : audit [INF] from='osd.2 [v2:192.168.123.108:6808/887940165,v1:192.168.123.108:6809/887940165]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.468783+0000 mon.vm03 (mon.0) 488 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.468783+0000 mon.vm03 (mon.0) 488 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: cluster 2026-03-06T21:18:17.473232+0000 mon.vm03 (mon.0) 489 : cluster [INF] osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367] boot 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: cluster 2026-03-06T21:18:17.473232+0000 mon.vm03 (mon.0) 489 : cluster [INF] osd.1 [v2:192.168.123.103:6802/213807367,v1:192.168.123.103:6803/213807367] boot 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: cluster 2026-03-06T21:18:17.473341+0000 mon.vm03 (mon.0) 490 : cluster [INF] osd.0 [v2:192.168.123.108:6800/956656304,v1:192.168.123.108:6801/956656304] boot 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: cluster 2026-03-06T21:18:17.473341+0000 mon.vm03 (mon.0) 490 : cluster [INF] osd.0 [v2:192.168.123.108:6800/956656304,v1:192.168.123.108:6801/956656304] boot 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: cluster 2026-03-06T21:18:17.473445+0000 mon.vm03 (mon.0) 491 : cluster [DBG] osdmap e16: 8 total, 2 up, 8 in 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: cluster 2026-03-06T21:18:17.473445+0000 mon.vm03 (mon.0) 491 : cluster [DBG] osdmap e16: 8 total, 2 up, 8 in 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.473828+0000 mon.vm03 (mon.0) 492 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.473828+0000 mon.vm03 (mon.0) 492 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.473969+0000 mon.vm03 (mon.0) 493 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.473969+0000 mon.vm03 (mon.0) 493 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.474243+0000 mon.vm03 (mon.0) 494 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.474243+0000 mon.vm03 (mon.0) 494 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.474379+0000 mon.vm03 (mon.0) 495 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.474379+0000 mon.vm03 (mon.0) 495 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.474745+0000 mon.vm03 (mon.0) 496 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.474745+0000 mon.vm03 (mon.0) 496 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.474851+0000 mon.vm03 (mon.0) 497 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.474851+0000 mon.vm03 (mon.0) 497 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.474898+0000 mon.vm03 (mon.0) 498 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.474898+0000 mon.vm03 (mon.0) 498 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.474939+0000 mon.vm03 (mon.0) 499 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.474939+0000 mon.vm03 (mon.0) 499 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.475137+0000 mon.vm03 (mon.0) 500 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.475137+0000 mon.vm03 (mon.0) 500 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.477189+0000 mon.vm08 (mon.1) 14 : audit [INF] from='osd.2 [v2:192.168.123.108:6808/887940165,v1:192.168.123.108:6809/887940165]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:17.770 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:17 vm03 bash[17055]: audit 2026-03-06T21:18:17.477189+0000 mon.vm08 (mon.1) 14 : audit [INF] from='osd.2 [v2:192.168.123.108:6808/887940165,v1:192.168.123.108:6809/887940165]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:18.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: cluster 2026-03-06T21:18:15.612535+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: cluster 2026-03-06T21:18:15.612535+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: cluster 2026-03-06T21:18:15.612615+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: cluster 2026-03-06T21:18:15.612615+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: cluster 2026-03-06T21:18:15.734583+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: cluster 2026-03-06T21:18:15.734583+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: cluster 2026-03-06T21:18:15.734639+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: cluster 2026-03-06T21:18:15.734639+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:17.585913+0000 mon.vm03 (mon.0) 501 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:17.585913+0000 mon.vm03 (mon.0) 501 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.471705+0000 mon.vm03 (mon.0) 502 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.471705+0000 mon.vm03 (mon.0) 502 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.471756+0000 mon.vm03 (mon.0) 503 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.471756+0000 mon.vm03 (mon.0) 503 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: cluster 2026-03-06T21:18:18.474208+0000 mon.vm03 (mon.0) 504 : cluster [DBG] osdmap e17: 8 total, 2 up, 8 in 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: cluster 2026-03-06T21:18:18.474208+0000 mon.vm03 (mon.0) 504 : cluster [DBG] osdmap e17: 8 total, 2 up, 8 in 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.474330+0000 mon.vm03 (mon.0) 505 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.474330+0000 mon.vm03 (mon.0) 505 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.474557+0000 mon.vm03 (mon.0) 506 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.474557+0000 mon.vm03 (mon.0) 506 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.474648+0000 mon.vm03 (mon.0) 507 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.474648+0000 mon.vm03 (mon.0) 507 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.474729+0000 mon.vm03 (mon.0) 508 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.474729+0000 mon.vm03 (mon.0) 508 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.474806+0000 mon.vm03 (mon.0) 509 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.474806+0000 mon.vm03 (mon.0) 509 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.474898+0000 mon.vm03 (mon.0) 510 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.474898+0000 mon.vm03 (mon.0) 510 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.474987+0000 mon.vm03 (mon.0) 511 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.474987+0000 mon.vm03 (mon.0) 511 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.477420+0000 mon.vm03 (mon.0) 512 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:18.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:18 vm08 bash[20828]: audit 2026-03-06T21:18:18.477420+0000 mon.vm03 (mon.0) 512 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:18.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: cluster 2026-03-06T21:18:15.612535+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: cluster 2026-03-06T21:18:15.612535+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: cluster 2026-03-06T21:18:15.612615+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: cluster 2026-03-06T21:18:15.612615+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: cluster 2026-03-06T21:18:15.734583+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: cluster 2026-03-06T21:18:15.734583+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: cluster 2026-03-06T21:18:15.734639+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: cluster 2026-03-06T21:18:15.734639+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:17.585913+0000 mon.vm03 (mon.0) 501 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:17.585913+0000 mon.vm03 (mon.0) 501 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.471705+0000 mon.vm03 (mon.0) 502 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.471705+0000 mon.vm03 (mon.0) 502 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.471756+0000 mon.vm03 (mon.0) 503 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.471756+0000 mon.vm03 (mon.0) 503 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: cluster 2026-03-06T21:18:18.474208+0000 mon.vm03 (mon.0) 504 : cluster [DBG] osdmap e17: 8 total, 2 up, 8 in 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: cluster 2026-03-06T21:18:18.474208+0000 mon.vm03 (mon.0) 504 : cluster [DBG] osdmap e17: 8 total, 2 up, 8 in 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.474330+0000 mon.vm03 (mon.0) 505 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.474330+0000 mon.vm03 (mon.0) 505 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.474557+0000 mon.vm03 (mon.0) 506 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.474557+0000 mon.vm03 (mon.0) 506 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.474648+0000 mon.vm03 (mon.0) 507 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.474648+0000 mon.vm03 (mon.0) 507 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.474729+0000 mon.vm03 (mon.0) 508 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.474729+0000 mon.vm03 (mon.0) 508 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.474806+0000 mon.vm03 (mon.0) 509 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.474806+0000 mon.vm03 (mon.0) 509 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.474898+0000 mon.vm03 (mon.0) 510 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.474898+0000 mon.vm03 (mon.0) 510 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.474987+0000 mon.vm03 (mon.0) 511 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.474987+0000 mon.vm03 (mon.0) 511 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.477420+0000 mon.vm03 (mon.0) 512 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:18.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:18 vm03 bash[17055]: audit 2026-03-06T21:18:18.477420+0000 mon.vm03 (mon.0) 512 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: cluster 2026-03-06T21:18:18.068737+0000 mgr.vm03.uwuzgl (mgr.14199) 91 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: cluster 2026-03-06T21:18:18.068737+0000 mgr.vm03.uwuzgl (mgr.14199) 91 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:18.955890+0000 mon.vm03 (mon.0) 513 : audit [INF] from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:18.955890+0000 mon.vm03 (mon.0) 513 : audit [INF] from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:18.960013+0000 mon.vm08 (mon.1) 15 : audit [INF] from='osd.4 [v2:192.168.123.108:6816/509534003,v1:192.168.123.108:6817/509534003]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:18.960013+0000 mon.vm08 (mon.1) 15 : audit [INF] from='osd.4 [v2:192.168.123.108:6816/509534003,v1:192.168.123.108:6817/509534003]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:18.960684+0000 mon.vm03 (mon.0) 514 : audit [INF] from='osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:18.960684+0000 mon.vm03 (mon.0) 514 : audit [INF] from='osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.474736+0000 mon.vm03 (mon.0) 515 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.474736+0000 mon.vm03 (mon.0) 515 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.474802+0000 mon.vm03 (mon.0) 516 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.474802+0000 mon.vm03 (mon.0) 516 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.474867+0000 mon.vm03 (mon.0) 517 : audit [INF] from='osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636]' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.474867+0000 mon.vm03 (mon.0) 517 : audit [INF] from='osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636]' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: cluster 2026-03-06T21:18:19.477515+0000 mon.vm03 (mon.0) 518 : cluster [INF] osd.2 [v2:192.168.123.108:6808/887940165,v1:192.168.123.108:6809/887940165] boot 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: cluster 2026-03-06T21:18:19.477515+0000 mon.vm03 (mon.0) 518 : cluster [INF] osd.2 [v2:192.168.123.108:6808/887940165,v1:192.168.123.108:6809/887940165] boot 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: cluster 2026-03-06T21:18:19.477623+0000 mon.vm03 (mon.0) 519 : cluster [DBG] osdmap e18: 8 total, 3 up, 8 in 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: cluster 2026-03-06T21:18:19.477623+0000 mon.vm03 (mon.0) 519 : cluster [DBG] osdmap e18: 8 total, 3 up, 8 in 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.478237+0000 mon.vm03 (mon.0) 520 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.478237+0000 mon.vm03 (mon.0) 520 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.478388+0000 mon.vm03 (mon.0) 521 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.478388+0000 mon.vm03 (mon.0) 521 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.478438+0000 mon.vm03 (mon.0) 522 : audit [INF] from='osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.478438+0000 mon.vm03 (mon.0) 522 : audit [INF] from='osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.478529+0000 mon.vm03 (mon.0) 523 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.478529+0000 mon.vm03 (mon.0) 523 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.478580+0000 mon.vm03 (mon.0) 524 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:19.832 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.478580+0000 mon.vm03 (mon.0) 524 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:19.833 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.478606+0000 mon.vm03 (mon.0) 525 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:19.833 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.478606+0000 mon.vm03 (mon.0) 525 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:19.833 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.478630+0000 mon.vm03 (mon.0) 526 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:19.833 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.478630+0000 mon.vm03 (mon.0) 526 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:19.833 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.480503+0000 mon.vm03 (mon.0) 527 : audit [INF] from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:19.833 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.480503+0000 mon.vm03 (mon.0) 527 : audit [INF] from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:19.833 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.482770+0000 mon.vm03 (mon.0) 528 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:19.833 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.482770+0000 mon.vm03 (mon.0) 528 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:19.833 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.483715+0000 mon.vm08 (mon.1) 16 : audit [INF] from='osd.4 [v2:192.168.123.108:6816/509534003,v1:192.168.123.108:6817/509534003]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:19.833 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:19 vm08 bash[20828]: audit 2026-03-06T21:18:19.483715+0000 mon.vm08 (mon.1) 16 : audit [INF] from='osd.4 [v2:192.168.123.108:6816/509534003,v1:192.168.123.108:6817/509534003]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:19.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: cluster 2026-03-06T21:18:18.068737+0000 mgr.vm03.uwuzgl (mgr.14199) 91 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T22:18:19.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: cluster 2026-03-06T21:18:18.068737+0000 mgr.vm03.uwuzgl (mgr.14199) 91 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:18.955890+0000 mon.vm03 (mon.0) 513 : audit [INF] from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:18.955890+0000 mon.vm03 (mon.0) 513 : audit [INF] from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:18.960013+0000 mon.vm08 (mon.1) 15 : audit [INF] from='osd.4 [v2:192.168.123.108:6816/509534003,v1:192.168.123.108:6817/509534003]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:18.960013+0000 mon.vm08 (mon.1) 15 : audit [INF] from='osd.4 [v2:192.168.123.108:6816/509534003,v1:192.168.123.108:6817/509534003]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:18.960684+0000 mon.vm03 (mon.0) 514 : audit [INF] from='osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:18.960684+0000 mon.vm03 (mon.0) 514 : audit [INF] from='osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.474736+0000 mon.vm03 (mon.0) 515 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.474736+0000 mon.vm03 (mon.0) 515 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.474802+0000 mon.vm03 (mon.0) 516 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.474802+0000 mon.vm03 (mon.0) 516 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.474867+0000 mon.vm03 (mon.0) 517 : audit [INF] from='osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636]' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.474867+0000 mon.vm03 (mon.0) 517 : audit [INF] from='osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636]' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: cluster 2026-03-06T21:18:19.477515+0000 mon.vm03 (mon.0) 518 : cluster [INF] osd.2 [v2:192.168.123.108:6808/887940165,v1:192.168.123.108:6809/887940165] boot 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: cluster 2026-03-06T21:18:19.477515+0000 mon.vm03 (mon.0) 518 : cluster [INF] osd.2 [v2:192.168.123.108:6808/887940165,v1:192.168.123.108:6809/887940165] boot 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: cluster 2026-03-06T21:18:19.477623+0000 mon.vm03 (mon.0) 519 : cluster [DBG] osdmap e18: 8 total, 3 up, 8 in 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: cluster 2026-03-06T21:18:19.477623+0000 mon.vm03 (mon.0) 519 : cluster [DBG] osdmap e18: 8 total, 3 up, 8 in 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.478237+0000 mon.vm03 (mon.0) 520 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.478237+0000 mon.vm03 (mon.0) 520 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.478388+0000 mon.vm03 (mon.0) 521 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.478388+0000 mon.vm03 (mon.0) 521 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.478438+0000 mon.vm03 (mon.0) 522 : audit [INF] from='osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.478438+0000 mon.vm03 (mon.0) 522 : audit [INF] from='osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.478529+0000 mon.vm03 (mon.0) 523 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.478529+0000 mon.vm03 (mon.0) 523 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.478580+0000 mon.vm03 (mon.0) 524 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.478580+0000 mon.vm03 (mon.0) 524 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.478606+0000 mon.vm03 (mon.0) 525 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.478606+0000 mon.vm03 (mon.0) 525 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.478630+0000 mon.vm03 (mon.0) 526 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.478630+0000 mon.vm03 (mon.0) 526 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.480503+0000 mon.vm03 (mon.0) 527 : audit [INF] from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.480503+0000 mon.vm03 (mon.0) 527 : audit [INF] from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.482770+0000 mon.vm03 (mon.0) 528 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.482770+0000 mon.vm03 (mon.0) 528 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.483715+0000 mon.vm08 (mon.1) 16 : audit [INF] from='osd.4 [v2:192.168.123.108:6816/509534003,v1:192.168.123.108:6817/509534003]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:19.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:19 vm03 bash[17055]: audit 2026-03-06T21:18:19.483715+0000 mon.vm08 (mon.1) 16 : audit [INF] from='osd.4 [v2:192.168.123.108:6816/509534003,v1:192.168.123.108:6817/509534003]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: cluster 2026-03-06T21:18:17.864433+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: cluster 2026-03-06T21:18:17.864433+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: cluster 2026-03-06T21:18:17.864495+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: cluster 2026-03-06T21:18:17.864495+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:19.829779+0000 mon.vm03 (mon.0) 529 : audit [INF] from='osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:19.829779+0000 mon.vm03 (mon.0) 529 : audit [INF] from='osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.070868+0000 mon.vm03 (mon.0) 530 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.070868+0000 mon.vm03 (mon.0) 530 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.416227+0000 mon.vm03 (mon.0) 531 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.416227+0000 mon.vm03 (mon.0) 531 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.465754+0000 mon.vm03 (mon.0) 532 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.465754+0000 mon.vm03 (mon.0) 532 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.477565+0000 mon.vm03 (mon.0) 533 : audit [INF] from='osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636]' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.477565+0000 mon.vm03 (mon.0) 533 : audit [INF] from='osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636]' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.477715+0000 mon.vm03 (mon.0) 534 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.477715+0000 mon.vm03 (mon.0) 534 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.477839+0000 mon.vm03 (mon.0) 535 : audit [INF] from='osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098]' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.477839+0000 mon.vm03 (mon.0) 535 : audit [INF] from='osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098]' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.477943+0000 mon.vm03 (mon.0) 536 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.477943+0000 mon.vm03 (mon.0) 536 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.478046+0000 mon.vm03 (mon.0) 537 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.478046+0000 mon.vm03 (mon.0) 537 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: cluster 2026-03-06T21:18:20.481278+0000 mon.vm03 (mon.0) 538 : cluster [INF] osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672] boot 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: cluster 2026-03-06T21:18:20.481278+0000 mon.vm03 (mon.0) 538 : cluster [INF] osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672] boot 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: cluster 2026-03-06T21:18:20.481432+0000 mon.vm03 (mon.0) 539 : cluster [DBG] osdmap e19: 8 total, 4 up, 8 in 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: cluster 2026-03-06T21:18:20.481432+0000 mon.vm03 (mon.0) 539 : cluster [DBG] osdmap e19: 8 total, 4 up, 8 in 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.482966+0000 mon.vm03 (mon.0) 540 : audit [INF] from='osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.482966+0000 mon.vm03 (mon.0) 540 : audit [INF] from='osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.483088+0000 mon.vm03 (mon.0) 541 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.483088+0000 mon.vm03 (mon.0) 541 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.483586+0000 mon.vm03 (mon.0) 542 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.483586+0000 mon.vm03 (mon.0) 542 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.483750+0000 mon.vm03 (mon.0) 543 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:20.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.483750+0000 mon.vm03 (mon.0) 543 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:20.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.483818+0000 mon.vm03 (mon.0) 544 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:20.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.483818+0000 mon.vm03 (mon.0) 544 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:20.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.483878+0000 mon.vm03 (mon.0) 545 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:20.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.483878+0000 mon.vm03 (mon.0) 545 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:20.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.483938+0000 mon.vm03 (mon.0) 546 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:20.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.483938+0000 mon.vm03 (mon.0) 546 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:20.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.484806+0000 mon.vm03 (mon.0) 547 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-06T22:18:20.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.484806+0000 mon.vm03 (mon.0) 547 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-06T22:18:20.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.488524+0000 mon.vm03 (mon.0) 548 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:20.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.488524+0000 mon.vm03 (mon.0) 548 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:20.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.490639+0000 mon.vm03 (mon.0) 549 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:20.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:20 vm03 bash[17055]: audit 2026-03-06T21:18:20.490639+0000 mon.vm03 (mon.0) 549 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:20.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: cluster 2026-03-06T21:18:17.864433+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: cluster 2026-03-06T21:18:17.864433+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: cluster 2026-03-06T21:18:17.864495+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: cluster 2026-03-06T21:18:17.864495+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:19.829779+0000 mon.vm03 (mon.0) 529 : audit [INF] from='osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:19.829779+0000 mon.vm03 (mon.0) 529 : audit [INF] from='osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.070868+0000 mon.vm03 (mon.0) 530 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.070868+0000 mon.vm03 (mon.0) 530 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.416227+0000 mon.vm03 (mon.0) 531 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.416227+0000 mon.vm03 (mon.0) 531 : audit [INF] from='osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672]' entity='osd.3' 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.465754+0000 mon.vm03 (mon.0) 532 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.465754+0000 mon.vm03 (mon.0) 532 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.477565+0000 mon.vm03 (mon.0) 533 : audit [INF] from='osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636]' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.477565+0000 mon.vm03 (mon.0) 533 : audit [INF] from='osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636]' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.477715+0000 mon.vm03 (mon.0) 534 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.477715+0000 mon.vm03 (mon.0) 534 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.477839+0000 mon.vm03 (mon.0) 535 : audit [INF] from='osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098]' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.477839+0000 mon.vm03 (mon.0) 535 : audit [INF] from='osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098]' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.477943+0000 mon.vm03 (mon.0) 536 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.477943+0000 mon.vm03 (mon.0) 536 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.478046+0000 mon.vm03 (mon.0) 537 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.478046+0000 mon.vm03 (mon.0) 537 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: cluster 2026-03-06T21:18:20.481278+0000 mon.vm03 (mon.0) 538 : cluster [INF] osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672] boot 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: cluster 2026-03-06T21:18:20.481278+0000 mon.vm03 (mon.0) 538 : cluster [INF] osd.3 [v2:192.168.123.103:6810/938793672,v1:192.168.123.103:6811/938793672] boot 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: cluster 2026-03-06T21:18:20.481432+0000 mon.vm03 (mon.0) 539 : cluster [DBG] osdmap e19: 8 total, 4 up, 8 in 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: cluster 2026-03-06T21:18:20.481432+0000 mon.vm03 (mon.0) 539 : cluster [DBG] osdmap e19: 8 total, 4 up, 8 in 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.482966+0000 mon.vm03 (mon.0) 540 : audit [INF] from='osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.482966+0000 mon.vm03 (mon.0) 540 : audit [INF] from='osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.483088+0000 mon.vm03 (mon.0) 541 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.483088+0000 mon.vm03 (mon.0) 541 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.483586+0000 mon.vm03 (mon.0) 542 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.483586+0000 mon.vm03 (mon.0) 542 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.483750+0000 mon.vm03 (mon.0) 543 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.483750+0000 mon.vm03 (mon.0) 543 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.483818+0000 mon.vm03 (mon.0) 544 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.483818+0000 mon.vm03 (mon.0) 544 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.483878+0000 mon.vm03 (mon.0) 545 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.483878+0000 mon.vm03 (mon.0) 545 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.483938+0000 mon.vm03 (mon.0) 546 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.483938+0000 mon.vm03 (mon.0) 546 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.484806+0000 mon.vm03 (mon.0) 547 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.484806+0000 mon.vm03 (mon.0) 547 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.488524+0000 mon.vm03 (mon.0) 548 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.488524+0000 mon.vm03 (mon.0) 548 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.490639+0000 mon.vm03 (mon.0) 549 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:20.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:20 vm08 bash[20828]: audit 2026-03-06T21:18:20.490639+0000 mon.vm03 (mon.0) 549 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: cluster 2026-03-06T21:18:18.604906+0000 osd.3 (osd.3) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: cluster 2026-03-06T21:18:18.604906+0000 osd.3 (osd.3) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: cluster 2026-03-06T21:18:18.604973+0000 osd.3 (osd.3) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: cluster 2026-03-06T21:18:18.604973+0000 osd.3 (osd.3) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: cluster 2026-03-06T21:18:20.069579+0000 mgr.vm03.uwuzgl (mgr.14199) 92 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: cluster 2026-03-06T21:18:20.069579+0000 mgr.vm03.uwuzgl (mgr.14199) 92 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: audit 2026-03-06T21:18:21.057013+0000 mon.vm03 (mon.0) 550 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: audit 2026-03-06T21:18:21.057013+0000 mon.vm03 (mon.0) 550 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: audit 2026-03-06T21:18:21.480803+0000 mon.vm03 (mon.0) 551 : audit [INF] from='osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098]' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: audit 2026-03-06T21:18:21.480803+0000 mon.vm03 (mon.0) 551 : audit [INF] from='osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098]' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: audit 2026-03-06T21:18:21.480885+0000 mon.vm03 (mon.0) 552 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: audit 2026-03-06T21:18:21.480885+0000 mon.vm03 (mon.0) 552 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: audit 2026-03-06T21:18:21.480915+0000 mon.vm03 (mon.0) 553 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: audit 2026-03-06T21:18:21.480915+0000 mon.vm03 (mon.0) 553 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: cluster 2026-03-06T21:18:21.484399+0000 mon.vm03 (mon.0) 554 : cluster [INF] osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636] boot 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: cluster 2026-03-06T21:18:21.484399+0000 mon.vm03 (mon.0) 554 : cluster [INF] osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636] boot 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: cluster 2026-03-06T21:18:21.484416+0000 mon.vm03 (mon.0) 555 : cluster [INF] osd.4 [v2:192.168.123.108:6816/509534003,v1:192.168.123.108:6817/509534003] boot 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: cluster 2026-03-06T21:18:21.484416+0000 mon.vm03 (mon.0) 555 : cluster [INF] osd.4 [v2:192.168.123.108:6816/509534003,v1:192.168.123.108:6817/509534003] boot 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: cluster 2026-03-06T21:18:21.484556+0000 mon.vm03 (mon.0) 556 : cluster [DBG] osdmap e20: 8 total, 6 up, 8 in 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: cluster 2026-03-06T21:18:21.484556+0000 mon.vm03 (mon.0) 556 : cluster [DBG] osdmap e20: 8 total, 6 up, 8 in 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: audit 2026-03-06T21:18:21.495106+0000 mon.vm03 (mon.0) 557 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: audit 2026-03-06T21:18:21.495106+0000 mon.vm03 (mon.0) 557 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: audit 2026-03-06T21:18:21.496008+0000 mon.vm03 (mon.0) 558 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: audit 2026-03-06T21:18:21.496008+0000 mon.vm03 (mon.0) 558 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: audit 2026-03-06T21:18:21.497149+0000 mon.vm03 (mon.0) 559 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: audit 2026-03-06T21:18:21.497149+0000 mon.vm03 (mon.0) 559 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: audit 2026-03-06T21:18:21.497481+0000 mon.vm03 (mon.0) 560 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:21.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:21 vm03 bash[17055]: audit 2026-03-06T21:18:21.497481+0000 mon.vm03 (mon.0) 560 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:22.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: cluster 2026-03-06T21:18:18.604906+0000 osd.3 (osd.3) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:22.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: cluster 2026-03-06T21:18:18.604906+0000 osd.3 (osd.3) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:22.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: cluster 2026-03-06T21:18:18.604973+0000 osd.3 (osd.3) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:22.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: cluster 2026-03-06T21:18:18.604973+0000 osd.3 (osd.3) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: cluster 2026-03-06T21:18:20.069579+0000 mgr.vm03.uwuzgl (mgr.14199) 92 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: cluster 2026-03-06T21:18:20.069579+0000 mgr.vm03.uwuzgl (mgr.14199) 92 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: audit 2026-03-06T21:18:21.057013+0000 mon.vm03 (mon.0) 550 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: audit 2026-03-06T21:18:21.057013+0000 mon.vm03 (mon.0) 550 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: audit 2026-03-06T21:18:21.480803+0000 mon.vm03 (mon.0) 551 : audit [INF] from='osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098]' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: audit 2026-03-06T21:18:21.480803+0000 mon.vm03 (mon.0) 551 : audit [INF] from='osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098]' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: audit 2026-03-06T21:18:21.480885+0000 mon.vm03 (mon.0) 552 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: audit 2026-03-06T21:18:21.480885+0000 mon.vm03 (mon.0) 552 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: audit 2026-03-06T21:18:21.480915+0000 mon.vm03 (mon.0) 553 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: audit 2026-03-06T21:18:21.480915+0000 mon.vm03 (mon.0) 553 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: cluster 2026-03-06T21:18:21.484399+0000 mon.vm03 (mon.0) 554 : cluster [INF] osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636] boot 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: cluster 2026-03-06T21:18:21.484399+0000 mon.vm03 (mon.0) 554 : cluster [INF] osd.5 [v2:192.168.123.103:6818/1425345636,v1:192.168.123.103:6819/1425345636] boot 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: cluster 2026-03-06T21:18:21.484416+0000 mon.vm03 (mon.0) 555 : cluster [INF] osd.4 [v2:192.168.123.108:6816/509534003,v1:192.168.123.108:6817/509534003] boot 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: cluster 2026-03-06T21:18:21.484416+0000 mon.vm03 (mon.0) 555 : cluster [INF] osd.4 [v2:192.168.123.108:6816/509534003,v1:192.168.123.108:6817/509534003] boot 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: cluster 2026-03-06T21:18:21.484556+0000 mon.vm03 (mon.0) 556 : cluster [DBG] osdmap e20: 8 total, 6 up, 8 in 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: cluster 2026-03-06T21:18:21.484556+0000 mon.vm03 (mon.0) 556 : cluster [DBG] osdmap e20: 8 total, 6 up, 8 in 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: audit 2026-03-06T21:18:21.495106+0000 mon.vm03 (mon.0) 557 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: audit 2026-03-06T21:18:21.495106+0000 mon.vm03 (mon.0) 557 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: audit 2026-03-06T21:18:21.496008+0000 mon.vm03 (mon.0) 558 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: audit 2026-03-06T21:18:21.496008+0000 mon.vm03 (mon.0) 558 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: audit 2026-03-06T21:18:21.497149+0000 mon.vm03 (mon.0) 559 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: audit 2026-03-06T21:18:21.497149+0000 mon.vm03 (mon.0) 559 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: audit 2026-03-06T21:18:21.497481+0000 mon.vm03 (mon.0) 560 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:22.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:21 vm08 bash[20828]: audit 2026-03-06T21:18:21.497481+0000 mon.vm03 (mon.0) 560 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:22.315 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: cluster 2026-03-06T21:18:19.909302+0000 osd.4 (osd.4) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: cluster 2026-03-06T21:18:19.909302+0000 osd.4 (osd.4) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: cluster 2026-03-06T21:18:19.909353+0000 osd.4 (osd.4) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: cluster 2026-03-06T21:18:19.909353+0000 osd.4 (osd.4) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: cluster 2026-03-06T21:18:19.923420+0000 osd.5 (osd.5) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: cluster 2026-03-06T21:18:19.923420+0000 osd.5 (osd.5) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: cluster 2026-03-06T21:18:19.923459+0000 osd.5 (osd.5) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: cluster 2026-03-06T21:18:19.923459+0000 osd.5 (osd.5) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: cluster 2026-03-06T21:18:22.490035+0000 mon.vm03 (mon.0) 561 : cluster [INF] osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098] boot 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: cluster 2026-03-06T21:18:22.490035+0000 mon.vm03 (mon.0) 561 : cluster [INF] osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098] boot 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: cluster 2026-03-06T21:18:22.490109+0000 mon.vm03 (mon.0) 562 : cluster [DBG] osdmap e21: 8 total, 7 up, 8 in 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: cluster 2026-03-06T21:18:22.490109+0000 mon.vm03 (mon.0) 562 : cluster [DBG] osdmap e21: 8 total, 7 up, 8 in 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: audit 2026-03-06T21:18:22.493772+0000 mon.vm03 (mon.0) 563 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: audit 2026-03-06T21:18:22.493772+0000 mon.vm03 (mon.0) 563 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: audit 2026-03-06T21:18:22.495243+0000 mon.vm03 (mon.0) 564 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: audit 2026-03-06T21:18:22.495243+0000 mon.vm03 (mon.0) 564 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: audit 2026-03-06T21:18:22.497467+0000 mon.vm03 (mon.0) 565 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: audit 2026-03-06T21:18:22.497467+0000 mon.vm03 (mon.0) 565 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: audit 2026-03-06T21:18:22.499463+0000 mon.vm03 (mon.0) 566 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: audit 2026-03-06T21:18:22.499463+0000 mon.vm03 (mon.0) 566 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: audit 2026-03-06T21:18:22.513135+0000 mon.vm03 (mon.0) 567 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' 2026-03-06T22:18:22.639 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:22 vm03 bash[17055]: audit 2026-03-06T21:18:22.513135+0000 mon.vm03 (mon.0) 567 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' 2026-03-06T22:18:22.816 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:18:22.905 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":21,"num_osds":8,"num_up_osds":7,"osd_up_since":1772831902,"num_in_osds":8,"osd_in_since":1772831879,"num_remapped_pgs":1} 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: cluster 2026-03-06T21:18:19.909302+0000 osd.4 (osd.4) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: cluster 2026-03-06T21:18:19.909302+0000 osd.4 (osd.4) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: cluster 2026-03-06T21:18:19.909353+0000 osd.4 (osd.4) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: cluster 2026-03-06T21:18:19.909353+0000 osd.4 (osd.4) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: cluster 2026-03-06T21:18:19.923420+0000 osd.5 (osd.5) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: cluster 2026-03-06T21:18:19.923420+0000 osd.5 (osd.5) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: cluster 2026-03-06T21:18:19.923459+0000 osd.5 (osd.5) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: cluster 2026-03-06T21:18:19.923459+0000 osd.5 (osd.5) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: cluster 2026-03-06T21:18:22.490035+0000 mon.vm03 (mon.0) 561 : cluster [INF] osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098] boot 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: cluster 2026-03-06T21:18:22.490035+0000 mon.vm03 (mon.0) 561 : cluster [INF] osd.6 [v2:192.168.123.108:6824/4262379098,v1:192.168.123.108:6825/4262379098] boot 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: cluster 2026-03-06T21:18:22.490109+0000 mon.vm03 (mon.0) 562 : cluster [DBG] osdmap e21: 8 total, 7 up, 8 in 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: cluster 2026-03-06T21:18:22.490109+0000 mon.vm03 (mon.0) 562 : cluster [DBG] osdmap e21: 8 total, 7 up, 8 in 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: audit 2026-03-06T21:18:22.493772+0000 mon.vm03 (mon.0) 563 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: audit 2026-03-06T21:18:22.493772+0000 mon.vm03 (mon.0) 563 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: audit 2026-03-06T21:18:22.495243+0000 mon.vm03 (mon.0) 564 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: audit 2026-03-06T21:18:22.495243+0000 mon.vm03 (mon.0) 564 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: audit 2026-03-06T21:18:22.497467+0000 mon.vm03 (mon.0) 565 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: audit 2026-03-06T21:18:22.497467+0000 mon.vm03 (mon.0) 565 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: audit 2026-03-06T21:18:22.499463+0000 mon.vm03 (mon.0) 566 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: audit 2026-03-06T21:18:22.499463+0000 mon.vm03 (mon.0) 566 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: audit 2026-03-06T21:18:22.513135+0000 mon.vm03 (mon.0) 567 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' 2026-03-06T22:18:23.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:22 vm08 bash[20828]: audit 2026-03-06T21:18:22.513135+0000 mon.vm03 (mon.0) 567 : audit [INF] from='osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929]' entity='osd.7' 2026-03-06T22:18:23.906 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd stat -f json 2026-03-06T22:18:23.912 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: cluster 2026-03-06T21:18:20.869432+0000 osd.6 (osd.6) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:23.912 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: cluster 2026-03-06T21:18:20.869432+0000 osd.6 (osd.6) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:23.912 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: cluster 2026-03-06T21:18:20.869501+0000 osd.6 (osd.6) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:23.912 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: cluster 2026-03-06T21:18:20.869501+0000 osd.6 (osd.6) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:23.912 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: cluster 2026-03-06T21:18:21.417278+0000 osd.7 (osd.7) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:23.912 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: cluster 2026-03-06T21:18:21.417278+0000 osd.7 (osd.7) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:23.913 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: cluster 2026-03-06T21:18:21.417331+0000 osd.7 (osd.7) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:23.913 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: cluster 2026-03-06T21:18:21.417331+0000 osd.7 (osd.7) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:23.913 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: cluster 2026-03-06T21:18:22.069803+0000 mgr.vm03.uwuzgl (mgr.14199) 93 : cluster [DBG] pgmap v46: 1 pgs: 1 unknown; 0 B data, 165 MiB used, 80 GiB / 80 GiB avail 2026-03-06T22:18:23.913 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: cluster 2026-03-06T21:18:22.069803+0000 mgr.vm03.uwuzgl (mgr.14199) 93 : cluster [DBG] pgmap v46: 1 pgs: 1 unknown; 0 B data, 165 MiB used, 80 GiB / 80 GiB avail 2026-03-06T22:18:23.913 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: audit 2026-03-06T21:18:22.810746+0000 mon.vm03 (mon.0) 568 : audit [DBG] from='client.? 192.168.123.103:0/3877693770' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:18:23.913 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: audit 2026-03-06T21:18:22.810746+0000 mon.vm03 (mon.0) 568 : audit [DBG] from='client.? 192.168.123.103:0/3877693770' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:18:23.913 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: cluster 2026-03-06T21:18:23.492407+0000 mon.vm03 (mon.0) 569 : cluster [INF] osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929] boot 2026-03-06T22:18:23.913 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: cluster 2026-03-06T21:18:23.492407+0000 mon.vm03 (mon.0) 569 : cluster [INF] osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929] boot 2026-03-06T22:18:23.913 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: cluster 2026-03-06T21:18:23.492592+0000 mon.vm03 (mon.0) 570 : cluster [DBG] osdmap e22: 8 total, 8 up, 8 in 2026-03-06T22:18:23.913 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: cluster 2026-03-06T21:18:23.492592+0000 mon.vm03 (mon.0) 570 : cluster [DBG] osdmap e22: 8 total, 8 up, 8 in 2026-03-06T22:18:23.913 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: audit 2026-03-06T21:18:23.493503+0000 mon.vm03 (mon.0) 571 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:23.913 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:23 vm03 bash[17055]: audit 2026-03-06T21:18:23.493503+0000 mon.vm03 (mon.0) 571 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: cluster 2026-03-06T21:18:20.869432+0000 osd.6 (osd.6) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: cluster 2026-03-06T21:18:20.869432+0000 osd.6 (osd.6) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: cluster 2026-03-06T21:18:20.869501+0000 osd.6 (osd.6) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: cluster 2026-03-06T21:18:20.869501+0000 osd.6 (osd.6) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: cluster 2026-03-06T21:18:21.417278+0000 osd.7 (osd.7) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: cluster 2026-03-06T21:18:21.417278+0000 osd.7 (osd.7) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: cluster 2026-03-06T21:18:21.417331+0000 osd.7 (osd.7) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: cluster 2026-03-06T21:18:21.417331+0000 osd.7 (osd.7) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: cluster 2026-03-06T21:18:22.069803+0000 mgr.vm03.uwuzgl (mgr.14199) 93 : cluster [DBG] pgmap v46: 1 pgs: 1 unknown; 0 B data, 165 MiB used, 80 GiB / 80 GiB avail 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: cluster 2026-03-06T21:18:22.069803+0000 mgr.vm03.uwuzgl (mgr.14199) 93 : cluster [DBG] pgmap v46: 1 pgs: 1 unknown; 0 B data, 165 MiB used, 80 GiB / 80 GiB avail 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: audit 2026-03-06T21:18:22.810746+0000 mon.vm03 (mon.0) 568 : audit [DBG] from='client.? 192.168.123.103:0/3877693770' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: audit 2026-03-06T21:18:22.810746+0000 mon.vm03 (mon.0) 568 : audit [DBG] from='client.? 192.168.123.103:0/3877693770' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: cluster 2026-03-06T21:18:23.492407+0000 mon.vm03 (mon.0) 569 : cluster [INF] osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929] boot 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: cluster 2026-03-06T21:18:23.492407+0000 mon.vm03 (mon.0) 569 : cluster [INF] osd.7 [v2:192.168.123.103:6826/2667958929,v1:192.168.123.103:6827/2667958929] boot 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: cluster 2026-03-06T21:18:23.492592+0000 mon.vm03 (mon.0) 570 : cluster [DBG] osdmap e22: 8 total, 8 up, 8 in 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: cluster 2026-03-06T21:18:23.492592+0000 mon.vm03 (mon.0) 570 : cluster [DBG] osdmap e22: 8 total, 8 up, 8 in 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: audit 2026-03-06T21:18:23.493503+0000 mon.vm03 (mon.0) 571 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:23.930 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:23 vm08 bash[20828]: audit 2026-03-06T21:18:23.493503+0000 mon.vm03 (mon.0) 571 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-06T22:18:25.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:24 vm03 bash[17055]: audit 2026-03-06T21:18:23.998070+0000 mon.vm03 (mon.0) 572 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:24 vm03 bash[17055]: audit 2026-03-06T21:18:23.998070+0000 mon.vm03 (mon.0) 572 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:24 vm03 bash[17055]: audit 2026-03-06T21:18:24.003735+0000 mon.vm03 (mon.0) 573 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:24 vm03 bash[17055]: audit 2026-03-06T21:18:24.003735+0000 mon.vm03 (mon.0) 573 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:24 vm03 bash[17055]: cluster 2026-03-06T21:18:24.070031+0000 mgr.vm03.uwuzgl (mgr.14199) 94 : cluster [DBG] pgmap v49: 1 pgs: 1 creating+remapped+peering; 0 B data, 671 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:24 vm03 bash[17055]: cluster 2026-03-06T21:18:24.070031+0000 mgr.vm03.uwuzgl (mgr.14199) 94 : cluster [DBG] pgmap v49: 1 pgs: 1 creating+remapped+peering; 0 B data, 671 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:24 vm03 bash[17055]: audit 2026-03-06T21:18:24.217826+0000 mon.vm03 (mon.0) 574 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:24 vm03 bash[17055]: audit 2026-03-06T21:18:24.217826+0000 mon.vm03 (mon.0) 574 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:24 vm03 bash[17055]: audit 2026-03-06T21:18:24.223836+0000 mon.vm03 (mon.0) 575 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:24 vm03 bash[17055]: audit 2026-03-06T21:18:24.223836+0000 mon.vm03 (mon.0) 575 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:24 vm03 bash[17055]: audit 2026-03-06T21:18:24.264545+0000 mon.vm03 (mon.0) 576 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:24 vm03 bash[17055]: audit 2026-03-06T21:18:24.264545+0000 mon.vm03 (mon.0) 576 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:24 vm03 bash[17055]: cluster 2026-03-06T21:18:24.494698+0000 mon.vm03 (mon.0) 577 : cluster [DBG] osdmap e23: 8 total, 8 up, 8 in 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:24 vm03 bash[17055]: cluster 2026-03-06T21:18:24.494698+0000 mon.vm03 (mon.0) 577 : cluster [DBG] osdmap e23: 8 total, 8 up, 8 in 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:25 vm03 bash[17055]: audit 2026-03-06T21:18:24.914618+0000 mon.vm03 (mon.0) 578 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:25 vm03 bash[17055]: audit 2026-03-06T21:18:24.914618+0000 mon.vm03 (mon.0) 578 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:25 vm03 bash[17055]: audit 2026-03-06T21:18:24.932043+0000 mon.vm03 (mon.0) 579 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:25 vm03 bash[17055]: audit 2026-03-06T21:18:24.932043+0000 mon.vm03 (mon.0) 579 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:25 vm03 bash[17055]: audit 2026-03-06T21:18:24.932306+0000 mon.vm03 (mon.0) 580 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:25 vm03 bash[17055]: audit 2026-03-06T21:18:24.932306+0000 mon.vm03 (mon.0) 580 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:25 vm03 bash[17055]: audit 2026-03-06T21:18:24.932391+0000 mon.vm03 (mon.0) 581 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:25 vm03 bash[17055]: audit 2026-03-06T21:18:24.932391+0000 mon.vm03 (mon.0) 581 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:25 vm03 bash[17055]: audit 2026-03-06T21:18:24.934919+0000 mon.vm03 (mon.0) 582 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:25 vm03 bash[17055]: audit 2026-03-06T21:18:24.934919+0000 mon.vm03 (mon.0) 582 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:25 vm03 bash[17055]: audit 2026-03-06T21:18:24.934980+0000 mon.vm03 (mon.0) 583 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:25 vm03 bash[17055]: audit 2026-03-06T21:18:24.934980+0000 mon.vm03 (mon.0) 583 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:25 vm03 bash[17055]: audit 2026-03-06T21:18:24.939216+0000 mon.vm08 (mon.1) 17 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:25 vm03 bash[17055]: audit 2026-03-06T21:18:24.939216+0000 mon.vm08 (mon.1) 17 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:25 vm03 bash[17055]: audit 2026-03-06T21:18:24.957506+0000 mon.vm08 (mon.1) 18 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T22:18:25.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:25 vm03 bash[17055]: audit 2026-03-06T21:18:24.957506+0000 mon.vm08 (mon.1) 18 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T22:18:25.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:23.998070+0000 mon.vm03 (mon.0) 572 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:25.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:23.998070+0000 mon.vm03 (mon.0) 572 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:25.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.003735+0000 mon.vm03 (mon.0) 573 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:25.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.003735+0000 mon.vm03 (mon.0) 573 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:25.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: cluster 2026-03-06T21:18:24.070031+0000 mgr.vm03.uwuzgl (mgr.14199) 94 : cluster [DBG] pgmap v49: 1 pgs: 1 creating+remapped+peering; 0 B data, 671 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: cluster 2026-03-06T21:18:24.070031+0000 mgr.vm03.uwuzgl (mgr.14199) 94 : cluster [DBG] pgmap v49: 1 pgs: 1 creating+remapped+peering; 0 B data, 671 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.217826+0000 mon.vm03 (mon.0) 574 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.217826+0000 mon.vm03 (mon.0) 574 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.223836+0000 mon.vm03 (mon.0) 575 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.223836+0000 mon.vm03 (mon.0) 575 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.264545+0000 mon.vm03 (mon.0) 576 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.264545+0000 mon.vm03 (mon.0) 576 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: cluster 2026-03-06T21:18:24.494698+0000 mon.vm03 (mon.0) 577 : cluster [DBG] osdmap e23: 8 total, 8 up, 8 in 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: cluster 2026-03-06T21:18:24.494698+0000 mon.vm03 (mon.0) 577 : cluster [DBG] osdmap e23: 8 total, 8 up, 8 in 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.914618+0000 mon.vm03 (mon.0) 578 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.914618+0000 mon.vm03 (mon.0) 578 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.932043+0000 mon.vm03 (mon.0) 579 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.932043+0000 mon.vm03 (mon.0) 579 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.932306+0000 mon.vm03 (mon.0) 580 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.932306+0000 mon.vm03 (mon.0) 580 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.932391+0000 mon.vm03 (mon.0) 581 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.932391+0000 mon.vm03 (mon.0) 581 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.934919+0000 mon.vm03 (mon.0) 582 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.934919+0000 mon.vm03 (mon.0) 582 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm03"}]: dispatch 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.934980+0000 mon.vm03 (mon.0) 583 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.934980+0000 mon.vm03 (mon.0) 583 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.939216+0000 mon.vm08 (mon.1) 17 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.939216+0000 mon.vm08 (mon.1) 17 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.957506+0000 mon.vm08 (mon.1) 18 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T22:18:25.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:25 vm08 bash[20828]: audit 2026-03-06T21:18:24.957506+0000 mon.vm08 (mon.1) 18 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-06T22:18:26.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:26 vm03 bash[17055]: cluster 2026-03-06T21:18:25.498004+0000 mon.vm03 (mon.0) 584 : cluster [DBG] osdmap e24: 8 total, 8 up, 8 in 2026-03-06T22:18:26.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:26 vm03 bash[17055]: cluster 2026-03-06T21:18:25.498004+0000 mon.vm03 (mon.0) 584 : cluster [DBG] osdmap e24: 8 total, 8 up, 8 in 2026-03-06T22:18:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:26 vm08 bash[20828]: cluster 2026-03-06T21:18:25.498004+0000 mon.vm03 (mon.0) 584 : cluster [DBG] osdmap e24: 8 total, 8 up, 8 in 2026-03-06T22:18:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:26 vm08 bash[20828]: cluster 2026-03-06T21:18:25.498004+0000 mon.vm03 (mon.0) 584 : cluster [DBG] osdmap e24: 8 total, 8 up, 8 in 2026-03-06T22:18:27.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:27 vm03 bash[17055]: cluster 2026-03-06T21:18:26.070301+0000 mgr.vm03.uwuzgl (mgr.14199) 95 : cluster [DBG] pgmap v52: 1 pgs: 1 creating+remapped+peering; 0 B data, 611 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:27.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:27 vm03 bash[17055]: cluster 2026-03-06T21:18:26.070301+0000 mgr.vm03.uwuzgl (mgr.14199) 95 : cluster [DBG] pgmap v52: 1 pgs: 1 creating+remapped+peering; 0 B data, 611 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:27 vm03 bash[17055]: cluster 2026-03-06T21:18:26.513922+0000 mon.vm03 (mon.0) 585 : cluster [DBG] mgrmap e19: vm03.uwuzgl(active, since 80s), standbys: vm08.tdhano 2026-03-06T22:18:27.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:27 vm03 bash[17055]: cluster 2026-03-06T21:18:26.513922+0000 mon.vm03 (mon.0) 585 : cluster [DBG] mgrmap e19: vm03.uwuzgl(active, since 80s), standbys: vm08.tdhano 2026-03-06T22:18:28.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:27 vm08 bash[20828]: cluster 2026-03-06T21:18:26.070301+0000 mgr.vm03.uwuzgl (mgr.14199) 95 : cluster [DBG] pgmap v52: 1 pgs: 1 creating+remapped+peering; 0 B data, 611 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:28.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:27 vm08 bash[20828]: cluster 2026-03-06T21:18:26.070301+0000 mgr.vm03.uwuzgl (mgr.14199) 95 : cluster [DBG] pgmap v52: 1 pgs: 1 creating+remapped+peering; 0 B data, 611 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:28.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:27 vm08 bash[20828]: cluster 2026-03-06T21:18:26.513922+0000 mon.vm03 (mon.0) 585 : cluster [DBG] mgrmap e19: vm03.uwuzgl(active, since 80s), standbys: vm08.tdhano 2026-03-06T22:18:28.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:27 vm08 bash[20828]: cluster 2026-03-06T21:18:26.513922+0000 mon.vm03 (mon.0) 585 : cluster [DBG] mgrmap e19: vm03.uwuzgl(active, since 80s), standbys: vm08.tdhano 2026-03-06T22:18:28.720 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:18:29.076 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:18:29.164 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":24,"num_osds":8,"num_up_osds":8,"osd_up_since":1772831903,"num_in_osds":8,"osd_in_since":1772831879,"num_remapped_pgs":0} 2026-03-06T22:18:29.165 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd dump --format=json 2026-03-06T22:18:29.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:29 vm03 bash[17055]: cluster 2026-03-06T21:18:28.070577+0000 mgr.vm03.uwuzgl (mgr.14199) 96 : cluster [DBG] pgmap v53: 1 pgs: 1 creating+remapped+peering; 0 B data, 612 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:29.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:29 vm03 bash[17055]: cluster 2026-03-06T21:18:28.070577+0000 mgr.vm03.uwuzgl (mgr.14199) 96 : cluster [DBG] pgmap v53: 1 pgs: 1 creating+remapped+peering; 0 B data, 612 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:29.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:29 vm03 bash[17055]: audit 2026-03-06T21:18:29.070544+0000 mon.vm03 (mon.0) 586 : audit [DBG] from='client.? 192.168.123.103:0/3005989588' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:18:29.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:29 vm03 bash[17055]: audit 2026-03-06T21:18:29.070544+0000 mon.vm03 (mon.0) 586 : audit [DBG] from='client.? 192.168.123.103:0/3005989588' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:18:29.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:29 vm03 bash[17055]: audit 2026-03-06T21:18:29.366907+0000 mon.vm03 (mon.0) 587 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:29.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:29 vm03 bash[17055]: audit 2026-03-06T21:18:29.366907+0000 mon.vm03 (mon.0) 587 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:29.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:29 vm03 bash[17055]: audit 2026-03-06T21:18:29.371425+0000 mon.vm03 (mon.0) 588 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:29.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:29 vm03 bash[17055]: audit 2026-03-06T21:18:29.371425+0000 mon.vm03 (mon.0) 588 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:29 vm08 bash[20828]: cluster 2026-03-06T21:18:28.070577+0000 mgr.vm03.uwuzgl (mgr.14199) 96 : cluster [DBG] pgmap v53: 1 pgs: 1 creating+remapped+peering; 0 B data, 612 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:29 vm08 bash[20828]: cluster 2026-03-06T21:18:28.070577+0000 mgr.vm03.uwuzgl (mgr.14199) 96 : cluster [DBG] pgmap v53: 1 pgs: 1 creating+remapped+peering; 0 B data, 612 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:29 vm08 bash[20828]: audit 2026-03-06T21:18:29.070544+0000 mon.vm03 (mon.0) 586 : audit [DBG] from='client.? 192.168.123.103:0/3005989588' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:18:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:29 vm08 bash[20828]: audit 2026-03-06T21:18:29.070544+0000 mon.vm03 (mon.0) 586 : audit [DBG] from='client.? 192.168.123.103:0/3005989588' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T22:18:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:29 vm08 bash[20828]: audit 2026-03-06T21:18:29.366907+0000 mon.vm03 (mon.0) 587 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:30.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:29 vm08 bash[20828]: audit 2026-03-06T21:18:29.366907+0000 mon.vm03 (mon.0) 587 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:30.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:29 vm08 bash[20828]: audit 2026-03-06T21:18:29.371425+0000 mon.vm03 (mon.0) 588 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:30.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:29 vm08 bash[20828]: audit 2026-03-06T21:18:29.371425+0000 mon.vm03 (mon.0) 588 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:30.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:30 vm03 bash[17055]: audit 2026-03-06T21:18:29.560739+0000 mon.vm03 (mon.0) 589 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:30.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:30 vm03 bash[17055]: audit 2026-03-06T21:18:29.560739+0000 mon.vm03 (mon.0) 589 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:30.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:30 vm03 bash[17055]: audit 2026-03-06T21:18:29.565756+0000 mon.vm03 (mon.0) 590 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:30.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:30 vm03 bash[17055]: audit 2026-03-06T21:18:29.565756+0000 mon.vm03 (mon.0) 590 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:31.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:30 vm08 bash[20828]: audit 2026-03-06T21:18:29.560739+0000 mon.vm03 (mon.0) 589 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:31.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:30 vm08 bash[20828]: audit 2026-03-06T21:18:29.560739+0000 mon.vm03 (mon.0) 589 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:31.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:30 vm08 bash[20828]: audit 2026-03-06T21:18:29.565756+0000 mon.vm03 (mon.0) 590 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:31.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:30 vm08 bash[20828]: audit 2026-03-06T21:18:29.565756+0000 mon.vm03 (mon.0) 590 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:31.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:31 vm03 bash[17055]: cluster 2026-03-06T21:18:30.070831+0000 mgr.vm03.uwuzgl (mgr.14199) 97 : cluster [DBG] pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:31.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:31 vm03 bash[17055]: cluster 2026-03-06T21:18:30.070831+0000 mgr.vm03.uwuzgl (mgr.14199) 97 : cluster [DBG] pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:32.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:31 vm08 bash[20828]: cluster 2026-03-06T21:18:30.070831+0000 mgr.vm03.uwuzgl (mgr.14199) 97 : cluster [DBG] pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:32.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:31 vm08 bash[20828]: cluster 2026-03-06T21:18:30.070831+0000 mgr.vm03.uwuzgl (mgr.14199) 97 : cluster [DBG] pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:32.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:32 vm03 bash[17055]: cluster 2026-03-06T21:18:32.071018+0000 mgr.vm03.uwuzgl (mgr.14199) 98 : cluster [DBG] pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:32.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:32 vm03 bash[17055]: cluster 2026-03-06T21:18:32.071018+0000 mgr.vm03.uwuzgl (mgr.14199) 98 : cluster [DBG] pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:33.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:32 vm08 bash[20828]: cluster 2026-03-06T21:18:32.071018+0000 mgr.vm03.uwuzgl (mgr.14199) 98 : cluster [DBG] pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:33.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:32 vm08 bash[20828]: cluster 2026-03-06T21:18:32.071018+0000 mgr.vm03.uwuzgl (mgr.14199) 98 : cluster [DBG] pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:34.103 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:18:34.533 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:18:34.533 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":24,"fsid":"894e000c-19a1-11f1-8dbe-23b24380a082","created":"2026-03-06T21:15:29.313521+0000","modified":"2026-03-06T21:18:25.492216+0000","last_up_change":"2026-03-06T21:18:23.485219+0000","last_in_change":"2026-03-06T21:17:59.430164+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":9,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-06T21:18:20.073606+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"24","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"198867f4-f98d-4ee8-a08e-17f3733bec4d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":956656304},{"type":"v1","addr":"192.168.123.108:6801","nonce":956656304}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":956656304},{"type":"v1","addr":"192.168.123.108:6803","nonce":956656304}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":956656304},{"type":"v1","addr":"192.168.123.108:6807","nonce":956656304}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":956656304},{"type":"v1","addr":"192.168.123.108:6805","nonce":956656304}]},"public_addr":"192.168.123.108:6801/956656304","cluster_addr":"192.168.123.108:6803/956656304","heartbeat_back_addr":"192.168.123.108:6807/956656304","heartbeat_front_addr":"192.168.123.108:6805/956656304","state":["exists","up"]},{"osd":1,"uuid":"b3b4d223-7698-4548-bde2-0d448c9b3714","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6802","nonce":213807367},{"type":"v1","addr":"192.168.123.103:6803","nonce":213807367}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6804","nonce":213807367},{"type":"v1","addr":"192.168.123.103:6805","nonce":213807367}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6808","nonce":213807367},{"type":"v1","addr":"192.168.123.103:6809","nonce":213807367}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6806","nonce":213807367},{"type":"v1","addr":"192.168.123.103:6807","nonce":213807367}]},"public_addr":"192.168.123.103:6803/213807367","cluster_addr":"192.168.123.103:6805/213807367","heartbeat_back_addr":"192.168.123.103:6809/213807367","heartbeat_front_addr":"192.168.123.103:6807/213807367","state":["exists","up"]},{"osd":2,"uuid":"a6b95b38-64b6-4e58-94c4-ae263e2dc1ff","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":18,"up_thru":19,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6808","nonce":887940165},{"type":"v1","addr":"192.168.123.108:6809","nonce":887940165}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6810","nonce":887940165},{"type":"v1","addr":"192.168.123.108:6811","nonce":887940165}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6814","nonce":887940165},{"type":"v1","addr":"192.168.123.108:6815","nonce":887940165}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6812","nonce":887940165},{"type":"v1","addr":"192.168.123.108:6813","nonce":887940165}]},"public_addr":"192.168.123.108:6809/887940165","cluster_addr":"192.168.123.108:6811/887940165","heartbeat_back_addr":"192.168.123.108:6815/887940165","heartbeat_front_addr":"192.168.123.108:6813/887940165","state":["exists","up"]},{"osd":3,"uuid":"8bb1dfca-f66a-4fa1-86a3-bb356250604f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6810","nonce":938793672},{"type":"v1","addr":"192.168.123.103:6811","nonce":938793672}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6812","nonce":938793672},{"type":"v1","addr":"192.168.123.103:6813","nonce":938793672}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6816","nonce":938793672},{"type":"v1","addr":"192.168.123.103:6817","nonce":938793672}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6814","nonce":938793672},{"type":"v1","addr":"192.168.123.103:6815","nonce":938793672}]},"public_addr":"192.168.123.103:6811/938793672","cluster_addr":"192.168.123.103:6813/938793672","heartbeat_back_addr":"192.168.123.103:6817/938793672","heartbeat_front_addr":"192.168.123.103:6815/938793672","state":["exists","up"]},{"osd":4,"uuid":"d184c72d-8b18-42f5-b09e-8d2a36892d30","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6816","nonce":509534003},{"type":"v1","addr":"192.168.123.108:6817","nonce":509534003}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6818","nonce":509534003},{"type":"v1","addr":"192.168.123.108:6819","nonce":509534003}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6822","nonce":509534003},{"type":"v1","addr":"192.168.123.108:6823","nonce":509534003}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6820","nonce":509534003},{"type":"v1","addr":"192.168.123.108:6821","nonce":509534003}]},"public_addr":"192.168.123.108:6817/509534003","cluster_addr":"192.168.123.108:6819/509534003","heartbeat_back_addr":"192.168.123.108:6823/509534003","heartbeat_front_addr":"192.168.123.108:6821/509534003","state":["exists","up"]},{"osd":5,"uuid":"58e80b7b-b647-452b-9f23-bf7ae1e3b6bb","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":21,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6818","nonce":1425345636},{"type":"v1","addr":"192.168.123.103:6819","nonce":1425345636}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6820","nonce":1425345636},{"type":"v1","addr":"192.168.123.103:6821","nonce":1425345636}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6824","nonce":1425345636},{"type":"v1","addr":"192.168.123.103:6825","nonce":1425345636}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6822","nonce":1425345636},{"type":"v1","addr":"192.168.123.103:6823","nonce":1425345636}]},"public_addr":"192.168.123.103:6819/1425345636","cluster_addr":"192.168.123.103:6821/1425345636","heartbeat_back_addr":"192.168.123.103:6825/1425345636","heartbeat_front_addr":"192.168.123.103:6823/1425345636","state":["exists","up"]},{"osd":6,"uuid":"bc2367fc-eb8e-4dec-9239-555bf4ed1271","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":22,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6824","nonce":4262379098},{"type":"v1","addr":"192.168.123.108:6825","nonce":4262379098}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6826","nonce":4262379098},{"type":"v1","addr":"192.168.123.108:6827","nonce":4262379098}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6830","nonce":4262379098},{"type":"v1","addr":"192.168.123.108:6831","nonce":4262379098}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6828","nonce":4262379098},{"type":"v1","addr":"192.168.123.108:6829","nonce":4262379098}]},"public_addr":"192.168.123.108:6825/4262379098","cluster_addr":"192.168.123.108:6827/4262379098","heartbeat_back_addr":"192.168.123.108:6831/4262379098","heartbeat_front_addr":"192.168.123.108:6829/4262379098","state":["exists","up"]},{"osd":7,"uuid":"4fa45945-b055-4d3b-b99b-47e0ac3cbe06","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":22,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6826","nonce":2667958929},{"type":"v1","addr":"192.168.123.103:6827","nonce":2667958929}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6828","nonce":2667958929},{"type":"v1","addr":"192.168.123.103:6829","nonce":2667958929}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6832","nonce":2667958929},{"type":"v1","addr":"192.168.123.103:6833","nonce":2667958929}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6830","nonce":2667958929},{"type":"v1","addr":"192.168.123.103:6831","nonce":2667958929}]},"public_addr":"192.168.123.103:6827/2667958929","cluster_addr":"192.168.123.103:6829/2667958929","heartbeat_back_addr":"192.168.123.103:6833/2667958929","heartbeat_front_addr":"192.168.123.103:6831/2667958929","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:15.734640+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:15.612617+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:17.864496+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:18.604975+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:19.909354+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:19.923460+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:20.869503+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:21.417333+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.103:0/3387502838":"2026-03-07T21:17:06.010844+0000","192.168.123.103:0/1377032643":"2026-03-07T21:17:06.010844+0000","192.168.123.103:0/4162486617":"2026-03-07T21:15:55.868997+0000","192.168.123.103:0/1021123444":"2026-03-07T21:15:55.868997+0000","192.168.123.103:0/1963453629":"2026-03-07T21:17:06.010844+0000","192.168.123.103:6800/750459842":"2026-03-07T21:15:55.868997+0000","192.168.123.103:6801/750459842":"2026-03-07T21:15:55.868997+0000","192.168.123.103:6800/1578443639":"2026-03-07T21:17:06.010844+0000","192.168.123.103:0/4023141877":"2026-03-07T21:15:55.868997+0000","192.168.123.103:0/352376855":"2026-03-07T21:16:17.853990+0000","192.168.123.103:0/19010882":"2026-03-07T21:16:17.853990+0000","192.168.123.103:0/499702899":"2026-03-07T21:16:17.853990+0000","192.168.123.103:6800/2711035263":"2026-03-07T21:16:17.853990+0000","192.168.123.103:6801/1578443639":"2026-03-07T21:17:06.010844+0000","192.168.123.103:6801/2711035263":"2026-03-07T21:16:17.853990+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-06T22:18:34.615 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-06T21:18:20.073606+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'is_stretch_pool': False, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '24', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}, 'read_balance': {'score_type': 'Fair distribution', 'score_acting': 7.889999866485596, 'score_stable': 7.889999866485596, 'optimal_score': 0.3799999952316284, 'raw_score_acting': 3, 'raw_score_stable': 3, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}}] 2026-03-06T22:18:34.615 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd pool get .mgr pg_num 2026-03-06T22:18:35.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:35 vm03 bash[17055]: cluster 2026-03-06T21:18:34.071304+0000 mgr.vm03.uwuzgl (mgr.14199) 99 : cluster [DBG] pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:35.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:35 vm03 bash[17055]: cluster 2026-03-06T21:18:34.071304+0000 mgr.vm03.uwuzgl (mgr.14199) 99 : cluster [DBG] pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:35.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:35 vm03 bash[17055]: audit 2026-03-06T21:18:34.526781+0000 mon.vm03 (mon.0) 591 : audit [DBG] from='client.? 192.168.123.103:0/3946154114' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T22:18:35.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:35 vm03 bash[17055]: audit 2026-03-06T21:18:34.526781+0000 mon.vm03 (mon.0) 591 : audit [DBG] from='client.? 192.168.123.103:0/3946154114' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T22:18:35.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:35 vm08 bash[20828]: cluster 2026-03-06T21:18:34.071304+0000 mgr.vm03.uwuzgl (mgr.14199) 99 : cluster [DBG] pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:35.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:35 vm08 bash[20828]: cluster 2026-03-06T21:18:34.071304+0000 mgr.vm03.uwuzgl (mgr.14199) 99 : cluster [DBG] pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:35.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:35 vm08 bash[20828]: audit 2026-03-06T21:18:34.526781+0000 mon.vm03 (mon.0) 591 : audit [DBG] from='client.? 192.168.123.103:0/3946154114' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T22:18:35.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:35 vm08 bash[20828]: audit 2026-03-06T21:18:34.526781+0000 mon.vm03 (mon.0) 591 : audit [DBG] from='client.? 192.168.123.103:0/3946154114' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T22:18:36.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: cephadm 2026-03-06T21:18:35.593014+0000 mgr.vm03.uwuzgl (mgr.14199) 100 : cephadm [INF] Detected new or changed devices on vm03 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: cephadm 2026-03-06T21:18:35.593014+0000 mgr.vm03.uwuzgl (mgr.14199) 100 : cephadm [INF] Detected new or changed devices on vm03 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.598890+0000 mon.vm03 (mon.0) 592 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.598890+0000 mon.vm03 (mon.0) 592 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.603434+0000 mon.vm03 (mon.0) 593 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.603434+0000 mon.vm03 (mon.0) 593 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.604193+0000 mon.vm03 (mon.0) 594 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.604193+0000 mon.vm03 (mon.0) 594 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: cephadm 2026-03-06T21:18:35.763062+0000 mgr.vm03.uwuzgl (mgr.14199) 101 : cephadm [INF] Detected new or changed devices on vm08 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: cephadm 2026-03-06T21:18:35.763062+0000 mgr.vm03.uwuzgl (mgr.14199) 101 : cephadm [INF] Detected new or changed devices on vm08 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.768899+0000 mon.vm03 (mon.0) 595 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.768899+0000 mon.vm03 (mon.0) 595 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.773056+0000 mon.vm03 (mon.0) 596 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.773056+0000 mon.vm03 (mon.0) 596 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.773869+0000 mon.vm03 (mon.0) 597 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.773869+0000 mon.vm03 (mon.0) 597 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.774460+0000 mon.vm03 (mon.0) 598 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.774460+0000 mon.vm03 (mon.0) 598 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.774865+0000 mon.vm03 (mon.0) 599 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.774865+0000 mon.vm03 (mon.0) 599 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.778333+0000 mon.vm03 (mon.0) 600 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.778333+0000 mon.vm03 (mon.0) 600 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.779855+0000 mon.vm03 (mon.0) 601 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:35.779855+0000 mon.vm03 (mon.0) 601 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:36.057372+0000 mon.vm03 (mon.0) 602 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: audit 2026-03-06T21:18:36.057372+0000 mon.vm03 (mon.0) 602 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: cluster 2026-03-06T21:18:36.072366+0000 mgr.vm03.uwuzgl (mgr.14199) 102 : cluster [DBG] pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:36.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:36 vm03 bash[17055]: cluster 2026-03-06T21:18:36.072366+0000 mgr.vm03.uwuzgl (mgr.14199) 102 : cluster [DBG] pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:37.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: cephadm 2026-03-06T21:18:35.593014+0000 mgr.vm03.uwuzgl (mgr.14199) 100 : cephadm [INF] Detected new or changed devices on vm03 2026-03-06T22:18:37.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: cephadm 2026-03-06T21:18:35.593014+0000 mgr.vm03.uwuzgl (mgr.14199) 100 : cephadm [INF] Detected new or changed devices on vm03 2026-03-06T22:18:37.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.598890+0000 mon.vm03 (mon.0) 592 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:37.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.598890+0000 mon.vm03 (mon.0) 592 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:37.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.603434+0000 mon.vm03 (mon.0) 593 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:37.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.603434+0000 mon.vm03 (mon.0) 593 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.604193+0000 mon.vm03 (mon.0) 594 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.604193+0000 mon.vm03 (mon.0) 594 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: cephadm 2026-03-06T21:18:35.763062+0000 mgr.vm03.uwuzgl (mgr.14199) 101 : cephadm [INF] Detected new or changed devices on vm08 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: cephadm 2026-03-06T21:18:35.763062+0000 mgr.vm03.uwuzgl (mgr.14199) 101 : cephadm [INF] Detected new or changed devices on vm08 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.768899+0000 mon.vm03 (mon.0) 595 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.768899+0000 mon.vm03 (mon.0) 595 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.773056+0000 mon.vm03 (mon.0) 596 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.773056+0000 mon.vm03 (mon.0) 596 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.773869+0000 mon.vm03 (mon.0) 597 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.773869+0000 mon.vm03 (mon.0) 597 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.774460+0000 mon.vm03 (mon.0) 598 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.774460+0000 mon.vm03 (mon.0) 598 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.774865+0000 mon.vm03 (mon.0) 599 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.774865+0000 mon.vm03 (mon.0) 599 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.778333+0000 mon.vm03 (mon.0) 600 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.778333+0000 mon.vm03 (mon.0) 600 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.779855+0000 mon.vm03 (mon.0) 601 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:35.779855+0000 mon.vm03 (mon.0) 601 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:36.057372+0000 mon.vm03 (mon.0) 602 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: audit 2026-03-06T21:18:36.057372+0000 mon.vm03 (mon.0) 602 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: cluster 2026-03-06T21:18:36.072366+0000 mgr.vm03.uwuzgl (mgr.14199) 102 : cluster [DBG] pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:37.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:36 vm08 bash[20828]: cluster 2026-03-06T21:18:36.072366+0000 mgr.vm03.uwuzgl (mgr.14199) 102 : cluster [DBG] pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:39.433 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:18:39.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:39 vm03 bash[17055]: cluster 2026-03-06T21:18:38.072656+0000 mgr.vm03.uwuzgl (mgr.14199) 103 : cluster [DBG] pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:39.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:39 vm03 bash[17055]: cluster 2026-03-06T21:18:38.072656+0000 mgr.vm03.uwuzgl (mgr.14199) 103 : cluster [DBG] pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:39.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:39 vm08 bash[20828]: cluster 2026-03-06T21:18:38.072656+0000 mgr.vm03.uwuzgl (mgr.14199) 103 : cluster [DBG] pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:39.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:39 vm08 bash[20828]: cluster 2026-03-06T21:18:38.072656+0000 mgr.vm03.uwuzgl (mgr.14199) 103 : cluster [DBG] pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:39.795 INFO:teuthology.orchestra.run.vm03.stdout:pg_num: 1 2026-03-06T22:18:39.875 INFO:tasks.cephadm:Setting up client nodes... 2026-03-06T22:18:39.875 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph auth get-or-create client.0 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-06T22:18:40.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:40 vm03 bash[17055]: audit 2026-03-06T21:18:39.793484+0000 mon.vm08 (mon.1) 19 : audit [DBG] from='client.? 192.168.123.103:0/3100358792' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-06T22:18:40.152 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:40 vm03 bash[17055]: audit 2026-03-06T21:18:39.793484+0000 mon.vm08 (mon.1) 19 : audit [DBG] from='client.? 192.168.123.103:0/3100358792' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-06T22:18:40.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:40 vm08 bash[20828]: audit 2026-03-06T21:18:39.793484+0000 mon.vm08 (mon.1) 19 : audit [DBG] from='client.? 192.168.123.103:0/3100358792' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-06T22:18:40.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:40 vm08 bash[20828]: audit 2026-03-06T21:18:39.793484+0000 mon.vm08 (mon.1) 19 : audit [DBG] from='client.? 192.168.123.103:0/3100358792' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-06T22:18:41.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:41 vm03 bash[17055]: cluster 2026-03-06T21:18:40.072963+0000 mgr.vm03.uwuzgl (mgr.14199) 104 : cluster [DBG] pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:41.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:41 vm03 bash[17055]: cluster 2026-03-06T21:18:40.072963+0000 mgr.vm03.uwuzgl (mgr.14199) 104 : cluster [DBG] pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:41.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:41 vm08 bash[20828]: cluster 2026-03-06T21:18:40.072963+0000 mgr.vm03.uwuzgl (mgr.14199) 104 : cluster [DBG] pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:41.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:41 vm08 bash[20828]: cluster 2026-03-06T21:18:40.072963+0000 mgr.vm03.uwuzgl (mgr.14199) 104 : cluster [DBG] pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:43.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:43 vm03 bash[17055]: cluster 2026-03-06T21:18:42.073223+0000 mgr.vm03.uwuzgl (mgr.14199) 105 : cluster [DBG] pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:43.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:43 vm03 bash[17055]: cluster 2026-03-06T21:18:42.073223+0000 mgr.vm03.uwuzgl (mgr.14199) 105 : cluster [DBG] pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:43.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:43 vm08 bash[20828]: cluster 2026-03-06T21:18:42.073223+0000 mgr.vm03.uwuzgl (mgr.14199) 105 : cluster [DBG] pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:43.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:43 vm08 bash[20828]: cluster 2026-03-06T21:18:42.073223+0000 mgr.vm03.uwuzgl (mgr.14199) 105 : cluster [DBG] pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:44.687 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:18:45.084 INFO:teuthology.orchestra.run.vm03.stdout:[client.0] 2026-03-06T22:18:45.084 INFO:teuthology.orchestra.run.vm03.stdout: key = AQC1RKtpLfogBBAAEyp1Qn1nV+BZDc/ooqG+fQ== 2026-03-06T22:18:45.159 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-06T22:18:45.159 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/etc/ceph/ceph.client.0.keyring 2026-03-06T22:18:45.159 DEBUG:teuthology.orchestra.run.vm03:> sudo chmod 0644 /etc/ceph/ceph.client.0.keyring 2026-03-06T22:18:45.173 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph auth get-or-create client.1 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-06T22:18:45.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:45 vm03 bash[17055]: cluster 2026-03-06T21:18:44.073490+0000 mgr.vm03.uwuzgl (mgr.14199) 106 : cluster [DBG] pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:45.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:45 vm03 bash[17055]: cluster 2026-03-06T21:18:44.073490+0000 mgr.vm03.uwuzgl (mgr.14199) 106 : cluster [DBG] pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:45.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:45 vm03 bash[17055]: audit 2026-03-06T21:18:45.069101+0000 mon.vm03 (mon.0) 603 : audit [INF] from='client.? 192.168.123.103:0/4116941191' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T22:18:45.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:45 vm03 bash[17055]: audit 2026-03-06T21:18:45.069101+0000 mon.vm03 (mon.0) 603 : audit [INF] from='client.? 192.168.123.103:0/4116941191' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T22:18:45.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:45 vm03 bash[17055]: audit 2026-03-06T21:18:45.073364+0000 mon.vm03 (mon.0) 604 : audit [INF] from='client.? 192.168.123.103:0/4116941191' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-06T22:18:45.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:45 vm03 bash[17055]: audit 2026-03-06T21:18:45.073364+0000 mon.vm03 (mon.0) 604 : audit [INF] from='client.? 192.168.123.103:0/4116941191' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-06T22:18:45.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:45 vm08 bash[20828]: cluster 2026-03-06T21:18:44.073490+0000 mgr.vm03.uwuzgl (mgr.14199) 106 : cluster [DBG] pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:45.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:45 vm08 bash[20828]: cluster 2026-03-06T21:18:44.073490+0000 mgr.vm03.uwuzgl (mgr.14199) 106 : cluster [DBG] pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:45.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:45 vm08 bash[20828]: audit 2026-03-06T21:18:45.069101+0000 mon.vm03 (mon.0) 603 : audit [INF] from='client.? 192.168.123.103:0/4116941191' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T22:18:45.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:45 vm08 bash[20828]: audit 2026-03-06T21:18:45.069101+0000 mon.vm03 (mon.0) 603 : audit [INF] from='client.? 192.168.123.103:0/4116941191' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T22:18:45.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:45 vm08 bash[20828]: audit 2026-03-06T21:18:45.073364+0000 mon.vm03 (mon.0) 604 : audit [INF] from='client.? 192.168.123.103:0/4116941191' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-06T22:18:45.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:45 vm08 bash[20828]: audit 2026-03-06T21:18:45.073364+0000 mon.vm03 (mon.0) 604 : audit [INF] from='client.? 192.168.123.103:0/4116941191' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-06T22:18:47.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:47 vm03 bash[17055]: cluster 2026-03-06T21:18:46.073692+0000 mgr.vm03.uwuzgl (mgr.14199) 107 : cluster [DBG] pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:47.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:47 vm03 bash[17055]: cluster 2026-03-06T21:18:46.073692+0000 mgr.vm03.uwuzgl (mgr.14199) 107 : cluster [DBG] pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:47.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:47 vm08 bash[20828]: cluster 2026-03-06T21:18:46.073692+0000 mgr.vm03.uwuzgl (mgr.14199) 107 : cluster [DBG] pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:47.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:47 vm08 bash[20828]: cluster 2026-03-06T21:18:46.073692+0000 mgr.vm03.uwuzgl (mgr.14199) 107 : cluster [DBG] pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:49.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:49 vm03 bash[17055]: cluster 2026-03-06T21:18:48.073916+0000 mgr.vm03.uwuzgl (mgr.14199) 108 : cluster [DBG] pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:49.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:49 vm03 bash[17055]: cluster 2026-03-06T21:18:48.073916+0000 mgr.vm03.uwuzgl (mgr.14199) 108 : cluster [DBG] pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:49.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:49 vm08 bash[20828]: cluster 2026-03-06T21:18:48.073916+0000 mgr.vm03.uwuzgl (mgr.14199) 108 : cluster [DBG] pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:49.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:49 vm08 bash[20828]: cluster 2026-03-06T21:18:48.073916+0000 mgr.vm03.uwuzgl (mgr.14199) 108 : cluster [DBG] pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:49.960 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm08/config 2026-03-06T22:18:50.361 INFO:teuthology.orchestra.run.vm08.stdout:[client.1] 2026-03-06T22:18:50.361 INFO:teuthology.orchestra.run.vm08.stdout: key = AQC6RKtpgfTvFBAAjSo6N8HbIIOMRrjmlg+eyQ== 2026-03-06T22:18:50.424 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-06T22:18:50.424 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/ceph.client.1.keyring 2026-03-06T22:18:50.424 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod 0644 /etc/ceph/ceph.client.1.keyring 2026-03-06T22:18:50.436 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-06T22:18:50.436 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-06T22:18:50.436 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph mgr dump --format=json 2026-03-06T22:18:51.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:51 vm03 bash[17055]: cluster 2026-03-06T21:18:50.074148+0000 mgr.vm03.uwuzgl (mgr.14199) 109 : cluster [DBG] pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:51.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:51 vm03 bash[17055]: cluster 2026-03-06T21:18:50.074148+0000 mgr.vm03.uwuzgl (mgr.14199) 109 : cluster [DBG] pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:51.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:51 vm03 bash[17055]: audit 2026-03-06T21:18:50.351189+0000 mon.vm03 (mon.0) 605 : audit [INF] from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T22:18:51.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:51 vm03 bash[17055]: audit 2026-03-06T21:18:50.351189+0000 mon.vm03 (mon.0) 605 : audit [INF] from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T22:18:51.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:51 vm03 bash[17055]: audit 2026-03-06T21:18:50.353699+0000 mon.vm03 (mon.0) 606 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-06T22:18:51.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:51 vm03 bash[17055]: audit 2026-03-06T21:18:50.353699+0000 mon.vm03 (mon.0) 606 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-06T22:18:51.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:51 vm03 bash[17055]: audit 2026-03-06T21:18:50.355393+0000 mon.vm08 (mon.1) 20 : audit [INF] from='client.? 192.168.123.108:0/1494471120' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T22:18:51.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:51 vm03 bash[17055]: audit 2026-03-06T21:18:50.355393+0000 mon.vm08 (mon.1) 20 : audit [INF] from='client.? 192.168.123.108:0/1494471120' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T22:18:51.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:51 vm03 bash[17055]: audit 2026-03-06T21:18:51.057509+0000 mon.vm03 (mon.0) 607 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:18:51.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:51 vm03 bash[17055]: audit 2026-03-06T21:18:51.057509+0000 mon.vm03 (mon.0) 607 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:18:51.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:51 vm08 bash[20828]: cluster 2026-03-06T21:18:50.074148+0000 mgr.vm03.uwuzgl (mgr.14199) 109 : cluster [DBG] pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:51.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:51 vm08 bash[20828]: cluster 2026-03-06T21:18:50.074148+0000 mgr.vm03.uwuzgl (mgr.14199) 109 : cluster [DBG] pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:51.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:51 vm08 bash[20828]: audit 2026-03-06T21:18:50.351189+0000 mon.vm03 (mon.0) 605 : audit [INF] from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T22:18:51.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:51 vm08 bash[20828]: audit 2026-03-06T21:18:50.351189+0000 mon.vm03 (mon.0) 605 : audit [INF] from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T22:18:51.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:51 vm08 bash[20828]: audit 2026-03-06T21:18:50.353699+0000 mon.vm03 (mon.0) 606 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-06T22:18:51.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:51 vm08 bash[20828]: audit 2026-03-06T21:18:50.353699+0000 mon.vm03 (mon.0) 606 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-06T22:18:51.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:51 vm08 bash[20828]: audit 2026-03-06T21:18:50.355393+0000 mon.vm08 (mon.1) 20 : audit [INF] from='client.? 192.168.123.108:0/1494471120' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T22:18:51.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:51 vm08 bash[20828]: audit 2026-03-06T21:18:50.355393+0000 mon.vm08 (mon.1) 20 : audit [INF] from='client.? 192.168.123.108:0/1494471120' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-06T22:18:51.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:51 vm08 bash[20828]: audit 2026-03-06T21:18:51.057509+0000 mon.vm03 (mon.0) 607 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:18:51.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:51 vm08 bash[20828]: audit 2026-03-06T21:18:51.057509+0000 mon.vm03 (mon.0) 607 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:18:53.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:53 vm03 bash[17055]: cluster 2026-03-06T21:18:52.074334+0000 mgr.vm03.uwuzgl (mgr.14199) 110 : cluster [DBG] pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:53.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:53 vm03 bash[17055]: cluster 2026-03-06T21:18:52.074334+0000 mgr.vm03.uwuzgl (mgr.14199) 110 : cluster [DBG] pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:53.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:53 vm08 bash[20828]: cluster 2026-03-06T21:18:52.074334+0000 mgr.vm03.uwuzgl (mgr.14199) 110 : cluster [DBG] pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:53.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:53 vm08 bash[20828]: cluster 2026-03-06T21:18:52.074334+0000 mgr.vm03.uwuzgl (mgr.14199) 110 : cluster [DBG] pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:55.226 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:18:55.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:55 vm08 bash[20828]: cluster 2026-03-06T21:18:54.074548+0000 mgr.vm03.uwuzgl (mgr.14199) 111 : cluster [DBG] pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:55.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:55 vm08 bash[20828]: cluster 2026-03-06T21:18:54.074548+0000 mgr.vm03.uwuzgl (mgr.14199) 111 : cluster [DBG] pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:55.598 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:18:55.609 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:55 vm03 bash[17055]: cluster 2026-03-06T21:18:54.074548+0000 mgr.vm03.uwuzgl (mgr.14199) 111 : cluster [DBG] pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:55.609 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:55 vm03 bash[17055]: cluster 2026-03-06T21:18:54.074548+0000 mgr.vm03.uwuzgl (mgr.14199) 111 : cluster [DBG] pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:55.659 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":19,"flags":0,"active_gid":14199,"active_name":"vm03.uwuzgl","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6800","nonce":2890947816},{"type":"v1","addr":"192.168.123.103:6801","nonce":2890947816}]},"active_addr":"192.168.123.103:6801/2890947816","active_change":"2026-03-06T21:17:06.011089+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":14214,"name":"vm08.tdhano","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","prometheus","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.103:8443/","prometheus":"http://192.168.123.103:9283/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":5,"active_clients":[{"name":"devicehealth","addrvec":[{"type":"v2","addr":"192.168.123.103:0","nonce":1055499200}]},{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.103:0","nonce":3505752254}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.103:0","nonce":613293069}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.103:0","nonce":651694206}]}]} 2026-03-06T22:18:55.661 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-06T22:18:55.661 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-06T22:18:55.661 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd dump --format=json 2026-03-06T22:18:56.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:56 vm08 bash[20828]: audit 2026-03-06T21:18:55.591708+0000 mon.vm03 (mon.0) 608 : audit [DBG] from='client.? 192.168.123.103:0/3966990595' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-06T22:18:56.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:56 vm08 bash[20828]: audit 2026-03-06T21:18:55.591708+0000 mon.vm03 (mon.0) 608 : audit [DBG] from='client.? 192.168.123.103:0/3966990595' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-06T22:18:56.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:56 vm03 bash[17055]: audit 2026-03-06T21:18:55.591708+0000 mon.vm03 (mon.0) 608 : audit [DBG] from='client.? 192.168.123.103:0/3966990595' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-06T22:18:56.514 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:56 vm03 bash[17055]: audit 2026-03-06T21:18:55.591708+0000 mon.vm03 (mon.0) 608 : audit [DBG] from='client.? 192.168.123.103:0/3966990595' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-06T22:18:57.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:57 vm08 bash[20828]: cluster 2026-03-06T21:18:56.074779+0000 mgr.vm03.uwuzgl (mgr.14199) 112 : cluster [DBG] pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:57.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:57 vm08 bash[20828]: cluster 2026-03-06T21:18:56.074779+0000 mgr.vm03.uwuzgl (mgr.14199) 112 : cluster [DBG] pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:57.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:57 vm03 bash[17055]: cluster 2026-03-06T21:18:56.074779+0000 mgr.vm03.uwuzgl (mgr.14199) 112 : cluster [DBG] pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:57.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:57 vm03 bash[17055]: cluster 2026-03-06T21:18:56.074779+0000 mgr.vm03.uwuzgl (mgr.14199) 112 : cluster [DBG] pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:59.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:59 vm08 bash[20828]: cluster 2026-03-06T21:18:58.075013+0000 mgr.vm03.uwuzgl (mgr.14199) 113 : cluster [DBG] pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:59.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:18:59 vm08 bash[20828]: cluster 2026-03-06T21:18:58.075013+0000 mgr.vm03.uwuzgl (mgr.14199) 113 : cluster [DBG] pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:59.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:59 vm03 bash[17055]: cluster 2026-03-06T21:18:58.075013+0000 mgr.vm03.uwuzgl (mgr.14199) 113 : cluster [DBG] pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:18:59.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:18:59 vm03 bash[17055]: cluster 2026-03-06T21:18:58.075013+0000 mgr.vm03.uwuzgl (mgr.14199) 113 : cluster [DBG] pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:00.441 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:00.814 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:19:00.814 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":24,"fsid":"894e000c-19a1-11f1-8dbe-23b24380a082","created":"2026-03-06T21:15:29.313521+0000","modified":"2026-03-06T21:18:25.492216+0000","last_up_change":"2026-03-06T21:18:23.485219+0000","last_in_change":"2026-03-06T21:17:59.430164+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":9,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-06T21:18:20.073606+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"24","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"198867f4-f98d-4ee8-a08e-17f3733bec4d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":956656304},{"type":"v1","addr":"192.168.123.108:6801","nonce":956656304}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":956656304},{"type":"v1","addr":"192.168.123.108:6803","nonce":956656304}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":956656304},{"type":"v1","addr":"192.168.123.108:6807","nonce":956656304}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":956656304},{"type":"v1","addr":"192.168.123.108:6805","nonce":956656304}]},"public_addr":"192.168.123.108:6801/956656304","cluster_addr":"192.168.123.108:6803/956656304","heartbeat_back_addr":"192.168.123.108:6807/956656304","heartbeat_front_addr":"192.168.123.108:6805/956656304","state":["exists","up"]},{"osd":1,"uuid":"b3b4d223-7698-4548-bde2-0d448c9b3714","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6802","nonce":213807367},{"type":"v1","addr":"192.168.123.103:6803","nonce":213807367}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6804","nonce":213807367},{"type":"v1","addr":"192.168.123.103:6805","nonce":213807367}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6808","nonce":213807367},{"type":"v1","addr":"192.168.123.103:6809","nonce":213807367}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6806","nonce":213807367},{"type":"v1","addr":"192.168.123.103:6807","nonce":213807367}]},"public_addr":"192.168.123.103:6803/213807367","cluster_addr":"192.168.123.103:6805/213807367","heartbeat_back_addr":"192.168.123.103:6809/213807367","heartbeat_front_addr":"192.168.123.103:6807/213807367","state":["exists","up"]},{"osd":2,"uuid":"a6b95b38-64b6-4e58-94c4-ae263e2dc1ff","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":18,"up_thru":19,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6808","nonce":887940165},{"type":"v1","addr":"192.168.123.108:6809","nonce":887940165}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6810","nonce":887940165},{"type":"v1","addr":"192.168.123.108:6811","nonce":887940165}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6814","nonce":887940165},{"type":"v1","addr":"192.168.123.108:6815","nonce":887940165}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6812","nonce":887940165},{"type":"v1","addr":"192.168.123.108:6813","nonce":887940165}]},"public_addr":"192.168.123.108:6809/887940165","cluster_addr":"192.168.123.108:6811/887940165","heartbeat_back_addr":"192.168.123.108:6815/887940165","heartbeat_front_addr":"192.168.123.108:6813/887940165","state":["exists","up"]},{"osd":3,"uuid":"8bb1dfca-f66a-4fa1-86a3-bb356250604f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6810","nonce":938793672},{"type":"v1","addr":"192.168.123.103:6811","nonce":938793672}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6812","nonce":938793672},{"type":"v1","addr":"192.168.123.103:6813","nonce":938793672}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6816","nonce":938793672},{"type":"v1","addr":"192.168.123.103:6817","nonce":938793672}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6814","nonce":938793672},{"type":"v1","addr":"192.168.123.103:6815","nonce":938793672}]},"public_addr":"192.168.123.103:6811/938793672","cluster_addr":"192.168.123.103:6813/938793672","heartbeat_back_addr":"192.168.123.103:6817/938793672","heartbeat_front_addr":"192.168.123.103:6815/938793672","state":["exists","up"]},{"osd":4,"uuid":"d184c72d-8b18-42f5-b09e-8d2a36892d30","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6816","nonce":509534003},{"type":"v1","addr":"192.168.123.108:6817","nonce":509534003}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6818","nonce":509534003},{"type":"v1","addr":"192.168.123.108:6819","nonce":509534003}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6822","nonce":509534003},{"type":"v1","addr":"192.168.123.108:6823","nonce":509534003}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6820","nonce":509534003},{"type":"v1","addr":"192.168.123.108:6821","nonce":509534003}]},"public_addr":"192.168.123.108:6817/509534003","cluster_addr":"192.168.123.108:6819/509534003","heartbeat_back_addr":"192.168.123.108:6823/509534003","heartbeat_front_addr":"192.168.123.108:6821/509534003","state":["exists","up"]},{"osd":5,"uuid":"58e80b7b-b647-452b-9f23-bf7ae1e3b6bb","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":21,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6818","nonce":1425345636},{"type":"v1","addr":"192.168.123.103:6819","nonce":1425345636}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6820","nonce":1425345636},{"type":"v1","addr":"192.168.123.103:6821","nonce":1425345636}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6824","nonce":1425345636},{"type":"v1","addr":"192.168.123.103:6825","nonce":1425345636}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6822","nonce":1425345636},{"type":"v1","addr":"192.168.123.103:6823","nonce":1425345636}]},"public_addr":"192.168.123.103:6819/1425345636","cluster_addr":"192.168.123.103:6821/1425345636","heartbeat_back_addr":"192.168.123.103:6825/1425345636","heartbeat_front_addr":"192.168.123.103:6823/1425345636","state":["exists","up"]},{"osd":6,"uuid":"bc2367fc-eb8e-4dec-9239-555bf4ed1271","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":22,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6824","nonce":4262379098},{"type":"v1","addr":"192.168.123.108:6825","nonce":4262379098}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6826","nonce":4262379098},{"type":"v1","addr":"192.168.123.108:6827","nonce":4262379098}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6830","nonce":4262379098},{"type":"v1","addr":"192.168.123.108:6831","nonce":4262379098}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6828","nonce":4262379098},{"type":"v1","addr":"192.168.123.108:6829","nonce":4262379098}]},"public_addr":"192.168.123.108:6825/4262379098","cluster_addr":"192.168.123.108:6827/4262379098","heartbeat_back_addr":"192.168.123.108:6831/4262379098","heartbeat_front_addr":"192.168.123.108:6829/4262379098","state":["exists","up"]},{"osd":7,"uuid":"4fa45945-b055-4d3b-b99b-47e0ac3cbe06","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":22,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6826","nonce":2667958929},{"type":"v1","addr":"192.168.123.103:6827","nonce":2667958929}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6828","nonce":2667958929},{"type":"v1","addr":"192.168.123.103:6829","nonce":2667958929}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6832","nonce":2667958929},{"type":"v1","addr":"192.168.123.103:6833","nonce":2667958929}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6830","nonce":2667958929},{"type":"v1","addr":"192.168.123.103:6831","nonce":2667958929}]},"public_addr":"192.168.123.103:6827/2667958929","cluster_addr":"192.168.123.103:6829/2667958929","heartbeat_back_addr":"192.168.123.103:6833/2667958929","heartbeat_front_addr":"192.168.123.103:6831/2667958929","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:15.734640+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:15.612617+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:17.864496+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:18.604975+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:19.909354+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:19.923460+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:20.869503+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:21.417333+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.103:0/3387502838":"2026-03-07T21:17:06.010844+0000","192.168.123.103:0/1377032643":"2026-03-07T21:17:06.010844+0000","192.168.123.103:0/4162486617":"2026-03-07T21:15:55.868997+0000","192.168.123.103:0/1021123444":"2026-03-07T21:15:55.868997+0000","192.168.123.103:0/1963453629":"2026-03-07T21:17:06.010844+0000","192.168.123.103:6800/750459842":"2026-03-07T21:15:55.868997+0000","192.168.123.103:6801/750459842":"2026-03-07T21:15:55.868997+0000","192.168.123.103:6800/1578443639":"2026-03-07T21:17:06.010844+0000","192.168.123.103:0/4023141877":"2026-03-07T21:15:55.868997+0000","192.168.123.103:0/352376855":"2026-03-07T21:16:17.853990+0000","192.168.123.103:0/19010882":"2026-03-07T21:16:17.853990+0000","192.168.123.103:0/499702899":"2026-03-07T21:16:17.853990+0000","192.168.123.103:6800/2711035263":"2026-03-07T21:16:17.853990+0000","192.168.123.103:6801/1578443639":"2026-03-07T21:17:06.010844+0000","192.168.123.103:6801/2711035263":"2026-03-07T21:16:17.853990+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-06T22:19:00.881 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-06T22:19:00.881 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd dump --format=json 2026-03-06T22:19:01.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:01 vm08 bash[20828]: cluster 2026-03-06T21:19:00.075255+0000 mgr.vm03.uwuzgl (mgr.14199) 114 : cluster [DBG] pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:01.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:01 vm08 bash[20828]: cluster 2026-03-06T21:19:00.075255+0000 mgr.vm03.uwuzgl (mgr.14199) 114 : cluster [DBG] pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:01.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:01 vm08 bash[20828]: audit 2026-03-06T21:19:00.808146+0000 mon.vm03 (mon.0) 609 : audit [DBG] from='client.? 192.168.123.103:0/3755961110' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T22:19:01.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:01 vm08 bash[20828]: audit 2026-03-06T21:19:00.808146+0000 mon.vm03 (mon.0) 609 : audit [DBG] from='client.? 192.168.123.103:0/3755961110' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T22:19:01.566 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:01 vm03 bash[17055]: cluster 2026-03-06T21:19:00.075255+0000 mgr.vm03.uwuzgl (mgr.14199) 114 : cluster [DBG] pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:01.566 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:01 vm03 bash[17055]: cluster 2026-03-06T21:19:00.075255+0000 mgr.vm03.uwuzgl (mgr.14199) 114 : cluster [DBG] pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:01.566 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:01 vm03 bash[17055]: audit 2026-03-06T21:19:00.808146+0000 mon.vm03 (mon.0) 609 : audit [DBG] from='client.? 192.168.123.103:0/3755961110' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T22:19:01.566 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:01 vm03 bash[17055]: audit 2026-03-06T21:19:00.808146+0000 mon.vm03 (mon.0) 609 : audit [DBG] from='client.? 192.168.123.103:0/3755961110' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T22:19:03.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:03 vm08 bash[20828]: cluster 2026-03-06T21:19:02.075506+0000 mgr.vm03.uwuzgl (mgr.14199) 115 : cluster [DBG] pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:03.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:03 vm08 bash[20828]: cluster 2026-03-06T21:19:02.075506+0000 mgr.vm03.uwuzgl (mgr.14199) 115 : cluster [DBG] pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:03.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:03 vm03 bash[17055]: cluster 2026-03-06T21:19:02.075506+0000 mgr.vm03.uwuzgl (mgr.14199) 115 : cluster [DBG] pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:03.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:03 vm03 bash[17055]: cluster 2026-03-06T21:19:02.075506+0000 mgr.vm03.uwuzgl (mgr.14199) 115 : cluster [DBG] pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:05.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:05 vm08 bash[20828]: cluster 2026-03-06T21:19:04.075732+0000 mgr.vm03.uwuzgl (mgr.14199) 116 : cluster [DBG] pgmap v71: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:05.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:05 vm08 bash[20828]: cluster 2026-03-06T21:19:04.075732+0000 mgr.vm03.uwuzgl (mgr.14199) 116 : cluster [DBG] pgmap v71: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:05.657 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:05.670 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:05 vm03 bash[17055]: cluster 2026-03-06T21:19:04.075732+0000 mgr.vm03.uwuzgl (mgr.14199) 116 : cluster [DBG] pgmap v71: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:05.670 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:05 vm03 bash[17055]: cluster 2026-03-06T21:19:04.075732+0000 mgr.vm03.uwuzgl (mgr.14199) 116 : cluster [DBG] pgmap v71: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:05.990 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:19:05.990 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":24,"fsid":"894e000c-19a1-11f1-8dbe-23b24380a082","created":"2026-03-06T21:15:29.313521+0000","modified":"2026-03-06T21:18:25.492216+0000","last_up_change":"2026-03-06T21:18:23.485219+0000","last_in_change":"2026-03-06T21:17:59.430164+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":9,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-06T21:18:20.073606+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"24","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"198867f4-f98d-4ee8-a08e-17f3733bec4d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":956656304},{"type":"v1","addr":"192.168.123.108:6801","nonce":956656304}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":956656304},{"type":"v1","addr":"192.168.123.108:6803","nonce":956656304}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":956656304},{"type":"v1","addr":"192.168.123.108:6807","nonce":956656304}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":956656304},{"type":"v1","addr":"192.168.123.108:6805","nonce":956656304}]},"public_addr":"192.168.123.108:6801/956656304","cluster_addr":"192.168.123.108:6803/956656304","heartbeat_back_addr":"192.168.123.108:6807/956656304","heartbeat_front_addr":"192.168.123.108:6805/956656304","state":["exists","up"]},{"osd":1,"uuid":"b3b4d223-7698-4548-bde2-0d448c9b3714","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6802","nonce":213807367},{"type":"v1","addr":"192.168.123.103:6803","nonce":213807367}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6804","nonce":213807367},{"type":"v1","addr":"192.168.123.103:6805","nonce":213807367}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6808","nonce":213807367},{"type":"v1","addr":"192.168.123.103:6809","nonce":213807367}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6806","nonce":213807367},{"type":"v1","addr":"192.168.123.103:6807","nonce":213807367}]},"public_addr":"192.168.123.103:6803/213807367","cluster_addr":"192.168.123.103:6805/213807367","heartbeat_back_addr":"192.168.123.103:6809/213807367","heartbeat_front_addr":"192.168.123.103:6807/213807367","state":["exists","up"]},{"osd":2,"uuid":"a6b95b38-64b6-4e58-94c4-ae263e2dc1ff","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":18,"up_thru":19,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6808","nonce":887940165},{"type":"v1","addr":"192.168.123.108:6809","nonce":887940165}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6810","nonce":887940165},{"type":"v1","addr":"192.168.123.108:6811","nonce":887940165}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6814","nonce":887940165},{"type":"v1","addr":"192.168.123.108:6815","nonce":887940165}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6812","nonce":887940165},{"type":"v1","addr":"192.168.123.108:6813","nonce":887940165}]},"public_addr":"192.168.123.108:6809/887940165","cluster_addr":"192.168.123.108:6811/887940165","heartbeat_back_addr":"192.168.123.108:6815/887940165","heartbeat_front_addr":"192.168.123.108:6813/887940165","state":["exists","up"]},{"osd":3,"uuid":"8bb1dfca-f66a-4fa1-86a3-bb356250604f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6810","nonce":938793672},{"type":"v1","addr":"192.168.123.103:6811","nonce":938793672}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6812","nonce":938793672},{"type":"v1","addr":"192.168.123.103:6813","nonce":938793672}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6816","nonce":938793672},{"type":"v1","addr":"192.168.123.103:6817","nonce":938793672}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6814","nonce":938793672},{"type":"v1","addr":"192.168.123.103:6815","nonce":938793672}]},"public_addr":"192.168.123.103:6811/938793672","cluster_addr":"192.168.123.103:6813/938793672","heartbeat_back_addr":"192.168.123.103:6817/938793672","heartbeat_front_addr":"192.168.123.103:6815/938793672","state":["exists","up"]},{"osd":4,"uuid":"d184c72d-8b18-42f5-b09e-8d2a36892d30","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6816","nonce":509534003},{"type":"v1","addr":"192.168.123.108:6817","nonce":509534003}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6818","nonce":509534003},{"type":"v1","addr":"192.168.123.108:6819","nonce":509534003}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6822","nonce":509534003},{"type":"v1","addr":"192.168.123.108:6823","nonce":509534003}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6820","nonce":509534003},{"type":"v1","addr":"192.168.123.108:6821","nonce":509534003}]},"public_addr":"192.168.123.108:6817/509534003","cluster_addr":"192.168.123.108:6819/509534003","heartbeat_back_addr":"192.168.123.108:6823/509534003","heartbeat_front_addr":"192.168.123.108:6821/509534003","state":["exists","up"]},{"osd":5,"uuid":"58e80b7b-b647-452b-9f23-bf7ae1e3b6bb","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":21,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6818","nonce":1425345636},{"type":"v1","addr":"192.168.123.103:6819","nonce":1425345636}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6820","nonce":1425345636},{"type":"v1","addr":"192.168.123.103:6821","nonce":1425345636}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6824","nonce":1425345636},{"type":"v1","addr":"192.168.123.103:6825","nonce":1425345636}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6822","nonce":1425345636},{"type":"v1","addr":"192.168.123.103:6823","nonce":1425345636}]},"public_addr":"192.168.123.103:6819/1425345636","cluster_addr":"192.168.123.103:6821/1425345636","heartbeat_back_addr":"192.168.123.103:6825/1425345636","heartbeat_front_addr":"192.168.123.103:6823/1425345636","state":["exists","up"]},{"osd":6,"uuid":"bc2367fc-eb8e-4dec-9239-555bf4ed1271","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":22,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6824","nonce":4262379098},{"type":"v1","addr":"192.168.123.108:6825","nonce":4262379098}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6826","nonce":4262379098},{"type":"v1","addr":"192.168.123.108:6827","nonce":4262379098}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6830","nonce":4262379098},{"type":"v1","addr":"192.168.123.108:6831","nonce":4262379098}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6828","nonce":4262379098},{"type":"v1","addr":"192.168.123.108:6829","nonce":4262379098}]},"public_addr":"192.168.123.108:6825/4262379098","cluster_addr":"192.168.123.108:6827/4262379098","heartbeat_back_addr":"192.168.123.108:6831/4262379098","heartbeat_front_addr":"192.168.123.108:6829/4262379098","state":["exists","up"]},{"osd":7,"uuid":"4fa45945-b055-4d3b-b99b-47e0ac3cbe06","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":22,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6826","nonce":2667958929},{"type":"v1","addr":"192.168.123.103:6827","nonce":2667958929}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6828","nonce":2667958929},{"type":"v1","addr":"192.168.123.103:6829","nonce":2667958929}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6832","nonce":2667958929},{"type":"v1","addr":"192.168.123.103:6833","nonce":2667958929}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6830","nonce":2667958929},{"type":"v1","addr":"192.168.123.103:6831","nonce":2667958929}]},"public_addr":"192.168.123.103:6827/2667958929","cluster_addr":"192.168.123.103:6829/2667958929","heartbeat_back_addr":"192.168.123.103:6833/2667958929","heartbeat_front_addr":"192.168.123.103:6831/2667958929","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:15.734640+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:15.612617+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:17.864496+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:18.604975+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:19.909354+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:19.923460+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:20.869503+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T21:18:21.417333+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.103:0/3387502838":"2026-03-07T21:17:06.010844+0000","192.168.123.103:0/1377032643":"2026-03-07T21:17:06.010844+0000","192.168.123.103:0/4162486617":"2026-03-07T21:15:55.868997+0000","192.168.123.103:0/1021123444":"2026-03-07T21:15:55.868997+0000","192.168.123.103:0/1963453629":"2026-03-07T21:17:06.010844+0000","192.168.123.103:6800/750459842":"2026-03-07T21:15:55.868997+0000","192.168.123.103:6801/750459842":"2026-03-07T21:15:55.868997+0000","192.168.123.103:6800/1578443639":"2026-03-07T21:17:06.010844+0000","192.168.123.103:0/4023141877":"2026-03-07T21:15:55.868997+0000","192.168.123.103:0/352376855":"2026-03-07T21:16:17.853990+0000","192.168.123.103:0/19010882":"2026-03-07T21:16:17.853990+0000","192.168.123.103:0/499702899":"2026-03-07T21:16:17.853990+0000","192.168.123.103:6800/2711035263":"2026-03-07T21:16:17.853990+0000","192.168.123.103:6801/1578443639":"2026-03-07T21:17:06.010844+0000","192.168.123.103:6801/2711035263":"2026-03-07T21:16:17.853990+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-06T22:19:06.051 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph tell osd.0 flush_pg_stats 2026-03-06T22:19:06.051 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph tell osd.1 flush_pg_stats 2026-03-06T22:19:06.052 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph tell osd.2 flush_pg_stats 2026-03-06T22:19:06.052 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph tell osd.3 flush_pg_stats 2026-03-06T22:19:06.052 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph tell osd.4 flush_pg_stats 2026-03-06T22:19:06.052 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph tell osd.5 flush_pg_stats 2026-03-06T22:19:06.052 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph tell osd.6 flush_pg_stats 2026-03-06T22:19:06.052 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph tell osd.7 flush_pg_stats 2026-03-06T22:19:06.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:06 vm03 bash[17055]: audit 2026-03-06T21:19:05.984252+0000 mon.vm03 (mon.0) 610 : audit [DBG] from='client.? 192.168.123.103:0/2899681757' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T22:19:06.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:06 vm03 bash[17055]: audit 2026-03-06T21:19:05.984252+0000 mon.vm03 (mon.0) 610 : audit [DBG] from='client.? 192.168.123.103:0/2899681757' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T22:19:06.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:06 vm03 bash[17055]: audit 2026-03-06T21:19:06.057558+0000 mon.vm03 (mon.0) 611 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:19:06.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:06 vm03 bash[17055]: audit 2026-03-06T21:19:06.057558+0000 mon.vm03 (mon.0) 611 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:19:06.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:06 vm08 bash[20828]: audit 2026-03-06T21:19:05.984252+0000 mon.vm03 (mon.0) 610 : audit [DBG] from='client.? 192.168.123.103:0/2899681757' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T22:19:06.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:06 vm08 bash[20828]: audit 2026-03-06T21:19:05.984252+0000 mon.vm03 (mon.0) 610 : audit [DBG] from='client.? 192.168.123.103:0/2899681757' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T22:19:06.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:06 vm08 bash[20828]: audit 2026-03-06T21:19:06.057558+0000 mon.vm03 (mon.0) 611 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:19:06.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:06 vm08 bash[20828]: audit 2026-03-06T21:19:06.057558+0000 mon.vm03 (mon.0) 611 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:19:07.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:07 vm08 bash[20828]: cluster 2026-03-06T21:19:06.075912+0000 mgr.vm03.uwuzgl (mgr.14199) 117 : cluster [DBG] pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:07.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:07 vm08 bash[20828]: cluster 2026-03-06T21:19:06.075912+0000 mgr.vm03.uwuzgl (mgr.14199) 117 : cluster [DBG] pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:07.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:07 vm03 bash[17055]: cluster 2026-03-06T21:19:06.075912+0000 mgr.vm03.uwuzgl (mgr.14199) 117 : cluster [DBG] pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:07.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:07 vm03 bash[17055]: cluster 2026-03-06T21:19:06.075912+0000 mgr.vm03.uwuzgl (mgr.14199) 117 : cluster [DBG] pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:09.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:09 vm08 bash[20828]: cluster 2026-03-06T21:19:08.076171+0000 mgr.vm03.uwuzgl (mgr.14199) 118 : cluster [DBG] pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:09.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:09 vm08 bash[20828]: cluster 2026-03-06T21:19:08.076171+0000 mgr.vm03.uwuzgl (mgr.14199) 118 : cluster [DBG] pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:09.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:09 vm03 bash[17055]: cluster 2026-03-06T21:19:08.076171+0000 mgr.vm03.uwuzgl (mgr.14199) 118 : cluster [DBG] pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:09.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:09 vm03 bash[17055]: cluster 2026-03-06T21:19:08.076171+0000 mgr.vm03.uwuzgl (mgr.14199) 118 : cluster [DBG] pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:11.634 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:11.635 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:11.638 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:11.639 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:11.640 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:11.640 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:11.642 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:11.642 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:11.660 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:11 vm03 bash[17055]: cluster 2026-03-06T21:19:10.076422+0000 mgr.vm03.uwuzgl (mgr.14199) 119 : cluster [DBG] pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:11.660 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:11 vm03 bash[17055]: cluster 2026-03-06T21:19:10.076422+0000 mgr.vm03.uwuzgl (mgr.14199) 119 : cluster [DBG] pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:11.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:11 vm08 bash[20828]: cluster 2026-03-06T21:19:10.076422+0000 mgr.vm03.uwuzgl (mgr.14199) 119 : cluster [DBG] pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:11.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:11 vm08 bash[20828]: cluster 2026-03-06T21:19:10.076422+0000 mgr.vm03.uwuzgl (mgr.14199) 119 : cluster [DBG] pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:12.638 INFO:teuthology.orchestra.run.vm03.stdout:68719476748 2026-03-06T22:19:12.638 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd last-stat-seq osd.0 2026-03-06T22:19:12.739 INFO:teuthology.orchestra.run.vm03.stdout:85899345931 2026-03-06T22:19:12.739 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd last-stat-seq osd.4 2026-03-06T22:19:12.987 INFO:teuthology.orchestra.run.vm03.stdout:90194313227 2026-03-06T22:19:12.987 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd last-stat-seq osd.6 2026-03-06T22:19:12.988 INFO:teuthology.orchestra.run.vm03.stdout:94489280523 2026-03-06T22:19:12.988 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd last-stat-seq osd.7 2026-03-06T22:19:13.068 INFO:teuthology.orchestra.run.vm03.stdout:81604378636 2026-03-06T22:19:13.068 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd last-stat-seq osd.3 2026-03-06T22:19:13.170 INFO:teuthology.orchestra.run.vm03.stdout:68719476748 2026-03-06T22:19:13.170 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd last-stat-seq osd.1 2026-03-06T22:19:13.178 INFO:teuthology.orchestra.run.vm03.stdout:77309411340 2026-03-06T22:19:13.178 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd last-stat-seq osd.2 2026-03-06T22:19:13.186 INFO:teuthology.orchestra.run.vm03.stdout:85899345932 2026-03-06T22:19:13.186 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph osd last-stat-seq osd.5 2026-03-06T22:19:13.659 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:13 vm03 bash[17055]: cluster 2026-03-06T21:19:12.076665+0000 mgr.vm03.uwuzgl (mgr.14199) 120 : cluster [DBG] pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:13.659 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:13 vm03 bash[17055]: cluster 2026-03-06T21:19:12.076665+0000 mgr.vm03.uwuzgl (mgr.14199) 120 : cluster [DBG] pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:13.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:13 vm08 bash[20828]: cluster 2026-03-06T21:19:12.076665+0000 mgr.vm03.uwuzgl (mgr.14199) 120 : cluster [DBG] pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:13.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:13 vm08 bash[20828]: cluster 2026-03-06T21:19:12.076665+0000 mgr.vm03.uwuzgl (mgr.14199) 120 : cluster [DBG] pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:15.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:15 vm03 bash[17055]: cluster 2026-03-06T21:19:14.076920+0000 mgr.vm03.uwuzgl (mgr.14199) 121 : cluster [DBG] pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:15.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:15 vm03 bash[17055]: cluster 2026-03-06T21:19:14.076920+0000 mgr.vm03.uwuzgl (mgr.14199) 121 : cluster [DBG] pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:15.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:15 vm08 bash[20828]: cluster 2026-03-06T21:19:14.076920+0000 mgr.vm03.uwuzgl (mgr.14199) 121 : cluster [DBG] pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:15.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:15 vm08 bash[20828]: cluster 2026-03-06T21:19:14.076920+0000 mgr.vm03.uwuzgl (mgr.14199) 121 : cluster [DBG] pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:17.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:17 vm03 bash[17055]: cluster 2026-03-06T21:19:16.077149+0000 mgr.vm03.uwuzgl (mgr.14199) 122 : cluster [DBG] pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:17.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:17 vm03 bash[17055]: cluster 2026-03-06T21:19:16.077149+0000 mgr.vm03.uwuzgl (mgr.14199) 122 : cluster [DBG] pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:17.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:17 vm08 bash[20828]: cluster 2026-03-06T21:19:16.077149+0000 mgr.vm03.uwuzgl (mgr.14199) 122 : cluster [DBG] pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:17.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:17 vm08 bash[20828]: cluster 2026-03-06T21:19:16.077149+0000 mgr.vm03.uwuzgl (mgr.14199) 122 : cluster [DBG] pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:17.995 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:17.996 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:17.998 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:17.998 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:18.001 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:18.001 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:18.003 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:18.005 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:19.090 INFO:teuthology.orchestra.run.vm03.stdout:68719476749 2026-03-06T22:19:19.250 INFO:tasks.cephadm.ceph_manager.ceph:need seq 68719476748 got 68719476749 for osd.1 2026-03-06T22:19:19.250 DEBUG:teuthology.parallel:result is None 2026-03-06T22:19:19.262 INFO:teuthology.orchestra.run.vm03.stdout:85899345933 2026-03-06T22:19:19.282 INFO:teuthology.orchestra.run.vm03.stdout:90194313228 2026-03-06T22:19:19.296 INFO:teuthology.orchestra.run.vm03.stdout:94489280524 2026-03-06T22:19:19.327 INFO:teuthology.orchestra.run.vm03.stdout:85899345933 2026-03-06T22:19:19.346 INFO:teuthology.orchestra.run.vm03.stdout:68719476750 2026-03-06T22:19:19.357 INFO:teuthology.orchestra.run.vm03.stdout:81604378637 2026-03-06T22:19:19.402 INFO:teuthology.orchestra.run.vm03.stdout:77309411341 2026-03-06T22:19:19.404 INFO:tasks.cephadm.ceph_manager.ceph:need seq 85899345931 got 85899345933 for osd.4 2026-03-06T22:19:19.404 DEBUG:teuthology.parallel:result is None 2026-03-06T22:19:19.404 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:19 vm03 bash[17055]: cluster 2026-03-06T21:19:18.077590+0000 mgr.vm03.uwuzgl (mgr.14199) 123 : cluster [DBG] pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:19.404 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:19 vm03 bash[17055]: cluster 2026-03-06T21:19:18.077590+0000 mgr.vm03.uwuzgl (mgr.14199) 123 : cluster [DBG] pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:19.404 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:19 vm03 bash[17055]: audit 2026-03-06T21:19:19.084760+0000 mon.vm03 (mon.0) 612 : audit [DBG] from='client.? 192.168.123.103:0/2185312116' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-06T22:19:19.404 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:19 vm03 bash[17055]: audit 2026-03-06T21:19:19.084760+0000 mon.vm03 (mon.0) 612 : audit [DBG] from='client.? 192.168.123.103:0/2185312116' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-06T22:19:19.404 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:19 vm03 bash[17055]: audit 2026-03-06T21:19:19.254511+0000 mon.vm03 (mon.0) 613 : audit [DBG] from='client.? 192.168.123.103:0/2967441143' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-06T22:19:19.405 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:19 vm03 bash[17055]: audit 2026-03-06T21:19:19.254511+0000 mon.vm03 (mon.0) 613 : audit [DBG] from='client.? 192.168.123.103:0/2967441143' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-06T22:19:19.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:19 vm08 bash[20828]: cluster 2026-03-06T21:19:18.077590+0000 mgr.vm03.uwuzgl (mgr.14199) 123 : cluster [DBG] pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:19.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:19 vm08 bash[20828]: cluster 2026-03-06T21:19:18.077590+0000 mgr.vm03.uwuzgl (mgr.14199) 123 : cluster [DBG] pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:19.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:19 vm08 bash[20828]: audit 2026-03-06T21:19:19.084760+0000 mon.vm03 (mon.0) 612 : audit [DBG] from='client.? 192.168.123.103:0/2185312116' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-06T22:19:19.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:19 vm08 bash[20828]: audit 2026-03-06T21:19:19.084760+0000 mon.vm03 (mon.0) 612 : audit [DBG] from='client.? 192.168.123.103:0/2185312116' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-06T22:19:19.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:19 vm08 bash[20828]: audit 2026-03-06T21:19:19.254511+0000 mon.vm03 (mon.0) 613 : audit [DBG] from='client.? 192.168.123.103:0/2967441143' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-06T22:19:19.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:19 vm08 bash[20828]: audit 2026-03-06T21:19:19.254511+0000 mon.vm03 (mon.0) 613 : audit [DBG] from='client.? 192.168.123.103:0/2967441143' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-06T22:19:19.503 INFO:tasks.cephadm.ceph_manager.ceph:need seq 90194313227 got 90194313228 for osd.6 2026-03-06T22:19:19.503 DEBUG:teuthology.parallel:result is None 2026-03-06T22:19:19.505 INFO:tasks.cephadm.ceph_manager.ceph:need seq 85899345932 got 85899345933 for osd.5 2026-03-06T22:19:19.505 DEBUG:teuthology.parallel:result is None 2026-03-06T22:19:19.506 INFO:tasks.cephadm.ceph_manager.ceph:need seq 94489280523 got 94489280524 for osd.7 2026-03-06T22:19:19.506 DEBUG:teuthology.parallel:result is None 2026-03-06T22:19:19.557 INFO:tasks.cephadm.ceph_manager.ceph:need seq 68719476748 got 68719476750 for osd.0 2026-03-06T22:19:19.557 DEBUG:teuthology.parallel:result is None 2026-03-06T22:19:19.567 INFO:tasks.cephadm.ceph_manager.ceph:need seq 81604378636 got 81604378637 for osd.3 2026-03-06T22:19:19.567 DEBUG:teuthology.parallel:result is None 2026-03-06T22:19:19.573 INFO:tasks.cephadm.ceph_manager.ceph:need seq 77309411340 got 77309411341 for osd.2 2026-03-06T22:19:19.573 DEBUG:teuthology.parallel:result is None 2026-03-06T22:19:19.573 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-06T22:19:19.573 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph pg dump --format=json 2026-03-06T22:19:20.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:20 vm03 bash[17055]: audit 2026-03-06T21:19:19.276414+0000 mon.vm03 (mon.0) 614 : audit [DBG] from='client.? 192.168.123.103:0/4022491843' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-06T22:19:20.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:20 vm03 bash[17055]: audit 2026-03-06T21:19:19.276414+0000 mon.vm03 (mon.0) 614 : audit [DBG] from='client.? 192.168.123.103:0/4022491843' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-06T22:19:20.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:20 vm03 bash[17055]: audit 2026-03-06T21:19:19.290043+0000 mon.vm03 (mon.0) 615 : audit [DBG] from='client.? 192.168.123.103:0/1822312334' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-06T22:19:20.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:20 vm03 bash[17055]: audit 2026-03-06T21:19:19.290043+0000 mon.vm03 (mon.0) 615 : audit [DBG] from='client.? 192.168.123.103:0/1822312334' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-06T22:19:20.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:20 vm03 bash[17055]: audit 2026-03-06T21:19:19.322287+0000 mon.vm03 (mon.0) 616 : audit [DBG] from='client.? 192.168.123.103:0/3278073530' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-06T22:19:20.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:20 vm03 bash[17055]: audit 2026-03-06T21:19:19.322287+0000 mon.vm03 (mon.0) 616 : audit [DBG] from='client.? 192.168.123.103:0/3278073530' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-06T22:19:20.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:20 vm03 bash[17055]: audit 2026-03-06T21:19:19.334220+0000 mon.vm03 (mon.0) 617 : audit [DBG] from='client.? 192.168.123.103:0/1627861916' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-06T22:19:20.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:20 vm03 bash[17055]: audit 2026-03-06T21:19:19.334220+0000 mon.vm03 (mon.0) 617 : audit [DBG] from='client.? 192.168.123.103:0/1627861916' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-06T22:19:20.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:20 vm03 bash[17055]: audit 2026-03-06T21:19:19.355590+0000 mon.vm08 (mon.1) 21 : audit [DBG] from='client.? 192.168.123.103:0/3080594849' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-06T22:19:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:20 vm03 bash[17055]: audit 2026-03-06T21:19:19.355590+0000 mon.vm08 (mon.1) 21 : audit [DBG] from='client.? 192.168.123.103:0/3080594849' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-06T22:19:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:20 vm03 bash[17055]: audit 2026-03-06T21:19:19.387386+0000 mon.vm03 (mon.0) 618 : audit [DBG] from='client.? 192.168.123.103:0/979961286' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-06T22:19:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:20 vm03 bash[17055]: audit 2026-03-06T21:19:19.387386+0000 mon.vm03 (mon.0) 618 : audit [DBG] from='client.? 192.168.123.103:0/979961286' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-06T22:19:20.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:20 vm08 bash[20828]: audit 2026-03-06T21:19:19.276414+0000 mon.vm03 (mon.0) 614 : audit [DBG] from='client.? 192.168.123.103:0/4022491843' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-06T22:19:20.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:20 vm08 bash[20828]: audit 2026-03-06T21:19:19.276414+0000 mon.vm03 (mon.0) 614 : audit [DBG] from='client.? 192.168.123.103:0/4022491843' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-06T22:19:20.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:20 vm08 bash[20828]: audit 2026-03-06T21:19:19.290043+0000 mon.vm03 (mon.0) 615 : audit [DBG] from='client.? 192.168.123.103:0/1822312334' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-06T22:19:20.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:20 vm08 bash[20828]: audit 2026-03-06T21:19:19.290043+0000 mon.vm03 (mon.0) 615 : audit [DBG] from='client.? 192.168.123.103:0/1822312334' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-06T22:19:20.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:20 vm08 bash[20828]: audit 2026-03-06T21:19:19.322287+0000 mon.vm03 (mon.0) 616 : audit [DBG] from='client.? 192.168.123.103:0/3278073530' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-06T22:19:20.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:20 vm08 bash[20828]: audit 2026-03-06T21:19:19.322287+0000 mon.vm03 (mon.0) 616 : audit [DBG] from='client.? 192.168.123.103:0/3278073530' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-06T22:19:20.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:20 vm08 bash[20828]: audit 2026-03-06T21:19:19.334220+0000 mon.vm03 (mon.0) 617 : audit [DBG] from='client.? 192.168.123.103:0/1627861916' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-06T22:19:20.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:20 vm08 bash[20828]: audit 2026-03-06T21:19:19.334220+0000 mon.vm03 (mon.0) 617 : audit [DBG] from='client.? 192.168.123.103:0/1627861916' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-06T22:19:20.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:20 vm08 bash[20828]: audit 2026-03-06T21:19:19.355590+0000 mon.vm08 (mon.1) 21 : audit [DBG] from='client.? 192.168.123.103:0/3080594849' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-06T22:19:20.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:20 vm08 bash[20828]: audit 2026-03-06T21:19:19.355590+0000 mon.vm08 (mon.1) 21 : audit [DBG] from='client.? 192.168.123.103:0/3080594849' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-06T22:19:20.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:20 vm08 bash[20828]: audit 2026-03-06T21:19:19.387386+0000 mon.vm03 (mon.0) 618 : audit [DBG] from='client.? 192.168.123.103:0/979961286' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-06T22:19:20.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:20 vm08 bash[20828]: audit 2026-03-06T21:19:19.387386+0000 mon.vm03 (mon.0) 618 : audit [DBG] from='client.? 192.168.123.103:0/979961286' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-06T22:19:21.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:21 vm03 bash[17055]: cluster 2026-03-06T21:19:20.077827+0000 mgr.vm03.uwuzgl (mgr.14199) 124 : cluster [DBG] pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:21.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:21 vm03 bash[17055]: cluster 2026-03-06T21:19:20.077827+0000 mgr.vm03.uwuzgl (mgr.14199) 124 : cluster [DBG] pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:21.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:21 vm03 bash[17055]: audit 2026-03-06T21:19:21.057787+0000 mon.vm03 (mon.0) 619 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:19:21.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:21 vm03 bash[17055]: audit 2026-03-06T21:19:21.057787+0000 mon.vm03 (mon.0) 619 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:19:21.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:21 vm08 bash[20828]: cluster 2026-03-06T21:19:20.077827+0000 mgr.vm03.uwuzgl (mgr.14199) 124 : cluster [DBG] pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:21.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:21 vm08 bash[20828]: cluster 2026-03-06T21:19:20.077827+0000 mgr.vm03.uwuzgl (mgr.14199) 124 : cluster [DBG] pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:21.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:21 vm08 bash[20828]: audit 2026-03-06T21:19:21.057787+0000 mon.vm03 (mon.0) 619 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:19:21.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:21 vm08 bash[20828]: audit 2026-03-06T21:19:21.057787+0000 mon.vm03 (mon.0) 619 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:19:23.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:23 vm03 bash[17055]: cluster 2026-03-06T21:19:22.078042+0000 mgr.vm03.uwuzgl (mgr.14199) 125 : cluster [DBG] pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:23.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:23 vm03 bash[17055]: cluster 2026-03-06T21:19:22.078042+0000 mgr.vm03.uwuzgl (mgr.14199) 125 : cluster [DBG] pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:23.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:23 vm08 bash[20828]: cluster 2026-03-06T21:19:22.078042+0000 mgr.vm03.uwuzgl (mgr.14199) 125 : cluster [DBG] pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:23.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:23 vm08 bash[20828]: cluster 2026-03-06T21:19:22.078042+0000 mgr.vm03.uwuzgl (mgr.14199) 125 : cluster [DBG] pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:24.394 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:24.720 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:19:24.720 INFO:teuthology.orchestra.run.vm03.stderr:dumped all 2026-03-06T22:19:24.796 INFO:teuthology.orchestra.run.vm03.stdout:{"pg_ready":true,"pg_map":{"version":81,"stamp":"2026-03-06T21:19:24.078156+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":3,"kb":167739392,"kb_used":627876,"kb_used_data":3148,"kb_used_omap":12,"kb_used_meta":214515,"kb_avail":167111516,"statfs":{"total":171765137408,"available":171122192384,"internally_reserved":0,"allocated":3223552,"data_stored":2048032,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":12711,"internal_metadata":219663961},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"12.001599"},"pg_stats":[{"pgid":"1.0","version":"23'32","reported_seq":57,"reported_epoch":24,"state":"active+clean","last_fresh":"2026-03-06T21:18:26.885076+0000","last_change":"2026-03-06T21:18:24.731575+0000","last_active":"2026-03-06T21:18:26.885076+0000","last_peered":"2026-03-06T21:18:26.885076+0000","last_clean":"2026-03-06T21:18:26.885076+0000","last_became_active":"2026-03-06T21:18:24.731453+0000","last_became_peered":"2026-03-06T21:18:24.731453+0000","last_unstale":"2026-03-06T21:18:26.885076+0000","last_undegraded":"2026-03-06T21:18:26.885076+0000","last_fullsized":"2026-03-06T21:18:26.885076+0000","mapping_epoch":22,"log_start":"0'0","ondisk_log_start":"0'0","created":19,"last_epoch_clean":23,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-06T21:18:20.474654+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-06T21:18:20.474654+0000","last_clean_scrub_stamp":"2026-03-06T21:18:20.474654+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-08T06:55:06.505795+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,5,2],"acting":[6,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":7,"up_from":22,"seq":94489280526,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":436664,"kb_used_data":224,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20530760,"statfs":{"total":21470642176,"available":21023498240,"internally_reserved":0,"allocated":229376,"data_stored":83774,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1584,"internal_metadata":27458000},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":21,"seq":90194313230,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27648,"kb_used_data":676,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939776,"statfs":{"total":21470642176,"available":21442330624,"internally_reserved":0,"allocated":692224,"data_stored":543054,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":20,"seq":85899345934,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27652,"kb_used_data":676,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939772,"statfs":{"total":21470642176,"available":21442326528,"internally_reserved":0,"allocated":692224,"data_stored":543054,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":20,"seq":85899345934,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27064,"kb_used_data":224,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940360,"statfs":{"total":21470642176,"available":21442928640,"internally_reserved":0,"allocated":229376,"data_stored":83774,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":19,"seq":81604378638,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27064,"kb_used_data":224,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940360,"statfs":{"total":21470642176,"available":21442928640,"internally_reserved":0,"allocated":229376,"data_stored":83774,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":18,"seq":77309411342,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27648,"kb_used_data":676,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939776,"statfs":{"total":21470642176,"available":21442330624,"internally_reserved":0,"allocated":692224,"data_stored":543054,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":16,"seq":68719476751,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27068,"kb_used_data":224,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940356,"statfs":{"total":21470642176,"available":21442924544,"internally_reserved":0,"allocated":229376,"data_stored":83774,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":16,"seq":68719476751,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27068,"kb_used_data":224,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940356,"statfs":{"total":21470642176,"available":21442924544,"internally_reserved":0,"allocated":229376,"data_stored":83774,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-06T22:19:24.796 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph pg dump --format=json 2026-03-06T22:19:25.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:25 vm03 bash[17055]: cluster 2026-03-06T21:19:24.078284+0000 mgr.vm03.uwuzgl (mgr.14199) 126 : cluster [DBG] pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:25.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:25 vm03 bash[17055]: cluster 2026-03-06T21:19:24.078284+0000 mgr.vm03.uwuzgl (mgr.14199) 126 : cluster [DBG] pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:25.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:25 vm08 bash[20828]: cluster 2026-03-06T21:19:24.078284+0000 mgr.vm03.uwuzgl (mgr.14199) 126 : cluster [DBG] pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:25.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:25 vm08 bash[20828]: cluster 2026-03-06T21:19:24.078284+0000 mgr.vm03.uwuzgl (mgr.14199) 126 : cluster [DBG] pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:26.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:26 vm03 bash[17055]: audit 2026-03-06T21:19:24.714602+0000 mgr.vm03.uwuzgl (mgr.14199) 127 : audit [DBG] from='client.14434 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:19:26.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:26 vm03 bash[17055]: audit 2026-03-06T21:19:24.714602+0000 mgr.vm03.uwuzgl (mgr.14199) 127 : audit [DBG] from='client.14434 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:19:26.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:26 vm08 bash[20828]: audit 2026-03-06T21:19:24.714602+0000 mgr.vm03.uwuzgl (mgr.14199) 127 : audit [DBG] from='client.14434 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:19:26.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:26 vm08 bash[20828]: audit 2026-03-06T21:19:24.714602+0000 mgr.vm03.uwuzgl (mgr.14199) 127 : audit [DBG] from='client.14434 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:19:27.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:27 vm03 bash[17055]: cluster 2026-03-06T21:19:26.078530+0000 mgr.vm03.uwuzgl (mgr.14199) 128 : cluster [DBG] pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:27.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:27 vm03 bash[17055]: cluster 2026-03-06T21:19:26.078530+0000 mgr.vm03.uwuzgl (mgr.14199) 128 : cluster [DBG] pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:27.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:27 vm08 bash[20828]: cluster 2026-03-06T21:19:26.078530+0000 mgr.vm03.uwuzgl (mgr.14199) 128 : cluster [DBG] pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:27.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:27 vm08 bash[20828]: cluster 2026-03-06T21:19:26.078530+0000 mgr.vm03.uwuzgl (mgr.14199) 128 : cluster [DBG] pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:29.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:29 vm08 bash[20828]: cluster 2026-03-06T21:19:28.078747+0000 mgr.vm03.uwuzgl (mgr.14199) 129 : cluster [DBG] pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:29.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:29 vm08 bash[20828]: cluster 2026-03-06T21:19:28.078747+0000 mgr.vm03.uwuzgl (mgr.14199) 129 : cluster [DBG] pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:29.570 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:29.582 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:29 vm03 bash[17055]: cluster 2026-03-06T21:19:28.078747+0000 mgr.vm03.uwuzgl (mgr.14199) 129 : cluster [DBG] pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:29.582 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:29 vm03 bash[17055]: cluster 2026-03-06T21:19:28.078747+0000 mgr.vm03.uwuzgl (mgr.14199) 129 : cluster [DBG] pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:29.949 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:19:29.949 INFO:teuthology.orchestra.run.vm03.stderr:dumped all 2026-03-06T22:19:30.010 INFO:teuthology.orchestra.run.vm03.stdout:{"pg_ready":true,"pg_map":{"version":83,"stamp":"2026-03-06T21:19:28.078654+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":3,"kb":167739392,"kb_used":627876,"kb_used_data":3148,"kb_used_omap":12,"kb_used_meta":214515,"kb_avail":167111516,"statfs":{"total":171765137408,"available":171122192384,"internally_reserved":0,"allocated":3223552,"data_stored":2048032,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":12711,"internal_metadata":219663961},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"12.001608"},"pg_stats":[{"pgid":"1.0","version":"23'32","reported_seq":57,"reported_epoch":24,"state":"active+clean","last_fresh":"2026-03-06T21:18:26.885076+0000","last_change":"2026-03-06T21:18:24.731575+0000","last_active":"2026-03-06T21:18:26.885076+0000","last_peered":"2026-03-06T21:18:26.885076+0000","last_clean":"2026-03-06T21:18:26.885076+0000","last_became_active":"2026-03-06T21:18:24.731453+0000","last_became_peered":"2026-03-06T21:18:24.731453+0000","last_unstale":"2026-03-06T21:18:26.885076+0000","last_undegraded":"2026-03-06T21:18:26.885076+0000","last_fullsized":"2026-03-06T21:18:26.885076+0000","mapping_epoch":22,"log_start":"0'0","ondisk_log_start":"0'0","created":19,"last_epoch_clean":23,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-06T21:18:20.474654+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-06T21:18:20.474654+0000","last_clean_scrub_stamp":"2026-03-06T21:18:20.474654+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-08T06:55:06.505795+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,5,2],"acting":[6,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":7,"up_from":22,"seq":94489280526,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":436664,"kb_used_data":224,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20530760,"statfs":{"total":21470642176,"available":21023498240,"internally_reserved":0,"allocated":229376,"data_stored":83774,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1584,"internal_metadata":27458000},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":21,"seq":90194313230,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27648,"kb_used_data":676,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939776,"statfs":{"total":21470642176,"available":21442330624,"internally_reserved":0,"allocated":692224,"data_stored":543054,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":20,"seq":85899345935,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27652,"kb_used_data":676,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939772,"statfs":{"total":21470642176,"available":21442326528,"internally_reserved":0,"allocated":692224,"data_stored":543054,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":20,"seq":85899345935,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27064,"kb_used_data":224,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940360,"statfs":{"total":21470642176,"available":21442928640,"internally_reserved":0,"allocated":229376,"data_stored":83774,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":19,"seq":81604378639,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27064,"kb_used_data":224,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940360,"statfs":{"total":21470642176,"available":21442928640,"internally_reserved":0,"allocated":229376,"data_stored":83774,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":18,"seq":77309411343,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27648,"kb_used_data":676,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939776,"statfs":{"total":21470642176,"available":21442330624,"internally_reserved":0,"allocated":692224,"data_stored":543054,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":16,"seq":68719476751,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27068,"kb_used_data":224,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940356,"statfs":{"total":21470642176,"available":21442924544,"internally_reserved":0,"allocated":229376,"data_stored":83774,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":16,"seq":68719476752,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27068,"kb_used_data":224,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940356,"statfs":{"total":21470642176,"available":21442924544,"internally_reserved":0,"allocated":229376,"data_stored":83774,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-06T22:19:30.011 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-06T22:19:30.011 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-06T22:19:30.011 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-06T22:19:30.011 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph health --format=json 2026-03-06T22:19:31.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:31 vm03 bash[17055]: audit 2026-03-06T21:19:29.943232+0000 mgr.vm03.uwuzgl (mgr.14199) 130 : audit [DBG] from='client.14438 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:19:31.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:31 vm03 bash[17055]: audit 2026-03-06T21:19:29.943232+0000 mgr.vm03.uwuzgl (mgr.14199) 130 : audit [DBG] from='client.14438 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:19:31.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:31 vm03 bash[17055]: cluster 2026-03-06T21:19:30.078955+0000 mgr.vm03.uwuzgl (mgr.14199) 131 : cluster [DBG] pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:31.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:31 vm03 bash[17055]: cluster 2026-03-06T21:19:30.078955+0000 mgr.vm03.uwuzgl (mgr.14199) 131 : cluster [DBG] pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:31.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:31 vm08 bash[20828]: audit 2026-03-06T21:19:29.943232+0000 mgr.vm03.uwuzgl (mgr.14199) 130 : audit [DBG] from='client.14438 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:19:31.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:31 vm08 bash[20828]: audit 2026-03-06T21:19:29.943232+0000 mgr.vm03.uwuzgl (mgr.14199) 130 : audit [DBG] from='client.14438 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:19:31.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:31 vm08 bash[20828]: cluster 2026-03-06T21:19:30.078955+0000 mgr.vm03.uwuzgl (mgr.14199) 131 : cluster [DBG] pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:31.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:31 vm08 bash[20828]: cluster 2026-03-06T21:19:30.078955+0000 mgr.vm03.uwuzgl (mgr.14199) 131 : cluster [DBG] pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:33.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:33 vm03 bash[17055]: cluster 2026-03-06T21:19:32.079167+0000 mgr.vm03.uwuzgl (mgr.14199) 132 : cluster [DBG] pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:33.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:33 vm03 bash[17055]: cluster 2026-03-06T21:19:32.079167+0000 mgr.vm03.uwuzgl (mgr.14199) 132 : cluster [DBG] pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:33.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:33 vm08 bash[20828]: cluster 2026-03-06T21:19:32.079167+0000 mgr.vm03.uwuzgl (mgr.14199) 132 : cluster [DBG] pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:33.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:33 vm08 bash[20828]: cluster 2026-03-06T21:19:32.079167+0000 mgr.vm03.uwuzgl (mgr.14199) 132 : cluster [DBG] pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:34.791 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:35.176 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:19:35.176 INFO:teuthology.orchestra.run.vm03.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-06T22:19:35.243 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-06T22:19:35.243 INFO:tasks.cephadm:Setup complete, yielding 2026-03-06T22:19:35.243 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-06T22:19:35.245 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm03.local 2026-03-06T22:19:35.245 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- bash -c 'ceph orch status' 2026-03-06T22:19:35.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:35 vm03 bash[17055]: cluster 2026-03-06T21:19:34.079413+0000 mgr.vm03.uwuzgl (mgr.14199) 133 : cluster [DBG] pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:35.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:35 vm03 bash[17055]: cluster 2026-03-06T21:19:34.079413+0000 mgr.vm03.uwuzgl (mgr.14199) 133 : cluster [DBG] pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:35.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:35 vm03 bash[17055]: audit 2026-03-06T21:19:35.170560+0000 mon.vm03 (mon.0) 620 : audit [DBG] from='client.? 192.168.123.103:0/3565266652' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-06T22:19:35.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:35 vm03 bash[17055]: audit 2026-03-06T21:19:35.170560+0000 mon.vm03 (mon.0) 620 : audit [DBG] from='client.? 192.168.123.103:0/3565266652' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-06T22:19:35.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:35 vm08 bash[20828]: cluster 2026-03-06T21:19:34.079413+0000 mgr.vm03.uwuzgl (mgr.14199) 133 : cluster [DBG] pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:35.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:35 vm08 bash[20828]: cluster 2026-03-06T21:19:34.079413+0000 mgr.vm03.uwuzgl (mgr.14199) 133 : cluster [DBG] pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:35.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:35 vm08 bash[20828]: audit 2026-03-06T21:19:35.170560+0000 mon.vm03 (mon.0) 620 : audit [DBG] from='client.? 192.168.123.103:0/3565266652' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-06T22:19:35.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:35 vm08 bash[20828]: audit 2026-03-06T21:19:35.170560+0000 mon.vm03 (mon.0) 620 : audit [DBG] from='client.? 192.168.123.103:0/3565266652' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-06T22:19:36.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:36 vm03 bash[17055]: audit 2026-03-06T21:19:35.819643+0000 mon.vm03 (mon.0) 621 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:19:36.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:36 vm03 bash[17055]: audit 2026-03-06T21:19:35.819643+0000 mon.vm03 (mon.0) 621 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:19:36.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:36 vm03 bash[17055]: audit 2026-03-06T21:19:36.058048+0000 mon.vm03 (mon.0) 622 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:19:36.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:36 vm03 bash[17055]: audit 2026-03-06T21:19:36.058048+0000 mon.vm03 (mon.0) 622 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:19:36.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:36 vm08 bash[20828]: audit 2026-03-06T21:19:35.819643+0000 mon.vm03 (mon.0) 621 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:19:36.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:36 vm08 bash[20828]: audit 2026-03-06T21:19:35.819643+0000 mon.vm03 (mon.0) 621 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:19:36.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:36 vm08 bash[20828]: audit 2026-03-06T21:19:36.058048+0000 mon.vm03 (mon.0) 622 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:19:36.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:36 vm08 bash[20828]: audit 2026-03-06T21:19:36.058048+0000 mon.vm03 (mon.0) 622 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:19:37.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:37 vm03 bash[17055]: cluster 2026-03-06T21:19:36.080184+0000 mgr.vm03.uwuzgl (mgr.14199) 134 : cluster [DBG] pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:37.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:37 vm03 bash[17055]: cluster 2026-03-06T21:19:36.080184+0000 mgr.vm03.uwuzgl (mgr.14199) 134 : cluster [DBG] pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:37.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:37 vm08 bash[20828]: cluster 2026-03-06T21:19:36.080184+0000 mgr.vm03.uwuzgl (mgr.14199) 134 : cluster [DBG] pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:37.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:37 vm08 bash[20828]: cluster 2026-03-06T21:19:36.080184+0000 mgr.vm03.uwuzgl (mgr.14199) 134 : cluster [DBG] pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:39.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:39 vm08 bash[20828]: cluster 2026-03-06T21:19:38.080433+0000 mgr.vm03.uwuzgl (mgr.14199) 135 : cluster [DBG] pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:39.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:39 vm08 bash[20828]: cluster 2026-03-06T21:19:38.080433+0000 mgr.vm03.uwuzgl (mgr.14199) 135 : cluster [DBG] pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:39.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:39 vm03 bash[17055]: cluster 2026-03-06T21:19:38.080433+0000 mgr.vm03.uwuzgl (mgr.14199) 135 : cluster [DBG] pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:39.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:39 vm03 bash[17055]: cluster 2026-03-06T21:19:38.080433+0000 mgr.vm03.uwuzgl (mgr.14199) 135 : cluster [DBG] pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:40.041 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:40.494 INFO:teuthology.orchestra.run.vm03.stdout:Backend: cephadm 2026-03-06T22:19:40.494 INFO:teuthology.orchestra.run.vm03.stdout:Available: Yes 2026-03-06T22:19:40.494 INFO:teuthology.orchestra.run.vm03.stdout:Paused: No 2026-03-06T22:19:40.583 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- bash -c 'ceph orch ps' 2026-03-06T22:19:41.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: cluster 2026-03-06T21:19:40.080744+0000 mgr.vm03.uwuzgl (mgr.14199) 136 : cluster [DBG] pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:41.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: cluster 2026-03-06T21:19:40.080744+0000 mgr.vm03.uwuzgl (mgr.14199) 136 : cluster [DBG] pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:41.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:40.486715+0000 mgr.vm03.uwuzgl (mgr.14199) 137 : audit [DBG] from='client.14446 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:19:41.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:40.486715+0000 mgr.vm03.uwuzgl (mgr.14199) 137 : audit [DBG] from='client.14446 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:19:41.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:40.809904+0000 mon.vm03 (mon.0) 623 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:40.809904+0000 mon.vm03 (mon.0) 623 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:40.816513+0000 mon.vm03 (mon.0) 624 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:40.816513+0000 mon.vm03 (mon.0) 624 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:40.920940+0000 mon.vm03 (mon.0) 625 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:40.920940+0000 mon.vm03 (mon.0) 625 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:40.927007+0000 mon.vm03 (mon.0) 626 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:40.927007+0000 mon.vm03 (mon.0) 626 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:41.237822+0000 mon.vm03 (mon.0) 627 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:19:41.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:41.237822+0000 mon.vm03 (mon.0) 627 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:19:41.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:41.238388+0000 mon.vm03 (mon.0) 628 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:19:41.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:41.238388+0000 mon.vm03 (mon.0) 628 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:19:41.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:41.243578+0000 mon.vm03 (mon.0) 629 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:41.243578+0000 mon.vm03 (mon.0) 629 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:41.245494+0000 mon.vm03 (mon.0) 630 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:19:41.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:41 vm03 bash[17055]: audit 2026-03-06T21:19:41.245494+0000 mon.vm03 (mon.0) 630 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: cluster 2026-03-06T21:19:40.080744+0000 mgr.vm03.uwuzgl (mgr.14199) 136 : cluster [DBG] pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: cluster 2026-03-06T21:19:40.080744+0000 mgr.vm03.uwuzgl (mgr.14199) 136 : cluster [DBG] pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:40.486715+0000 mgr.vm03.uwuzgl (mgr.14199) 137 : audit [DBG] from='client.14446 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:40.486715+0000 mgr.vm03.uwuzgl (mgr.14199) 137 : audit [DBG] from='client.14446 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:40.809904+0000 mon.vm03 (mon.0) 623 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:40.809904+0000 mon.vm03 (mon.0) 623 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:40.816513+0000 mon.vm03 (mon.0) 624 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:40.816513+0000 mon.vm03 (mon.0) 624 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:40.920940+0000 mon.vm03 (mon.0) 625 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:40.920940+0000 mon.vm03 (mon.0) 625 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:40.927007+0000 mon.vm03 (mon.0) 626 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:40.927007+0000 mon.vm03 (mon.0) 626 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:41.237822+0000 mon.vm03 (mon.0) 627 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:41.237822+0000 mon.vm03 (mon.0) 627 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:41.238388+0000 mon.vm03 (mon.0) 628 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:41.238388+0000 mon.vm03 (mon.0) 628 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:41.243578+0000 mon.vm03 (mon.0) 629 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:41.243578+0000 mon.vm03 (mon.0) 629 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:41.245494+0000 mon.vm03 (mon.0) 630 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:19:41.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:41 vm08 bash[20828]: audit 2026-03-06T21:19:41.245494+0000 mon.vm03 (mon.0) 630 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:19:43.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:43 vm03 bash[17055]: cluster 2026-03-06T21:19:42.081013+0000 mgr.vm03.uwuzgl (mgr.14199) 138 : cluster [DBG] pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:43.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:43 vm03 bash[17055]: cluster 2026-03-06T21:19:42.081013+0000 mgr.vm03.uwuzgl (mgr.14199) 138 : cluster [DBG] pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:43.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:43 vm08 bash[20828]: cluster 2026-03-06T21:19:42.081013+0000 mgr.vm03.uwuzgl (mgr.14199) 138 : cluster [DBG] pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:43.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:43 vm08 bash[20828]: cluster 2026-03-06T21:19:42.081013+0000 mgr.vm03.uwuzgl (mgr.14199) 138 : cluster [DBG] pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:45.448 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:45.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:45 vm03 bash[17055]: cluster 2026-03-06T21:19:44.081303+0000 mgr.vm03.uwuzgl (mgr.14199) 139 : cluster [DBG] pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:45.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:45 vm03 bash[17055]: cluster 2026-03-06T21:19:44.081303+0000 mgr.vm03.uwuzgl (mgr.14199) 139 : cluster [DBG] pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:45.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:45 vm08 bash[20828]: cluster 2026-03-06T21:19:44.081303+0000 mgr.vm03.uwuzgl (mgr.14199) 139 : cluster [DBG] pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:45.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:45 vm08 bash[20828]: cluster 2026-03-06T21:19:44.081303+0000 mgr.vm03.uwuzgl (mgr.14199) 139 : cluster [DBG] pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:45.828 INFO:teuthology.orchestra.run.vm03.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-06T22:19:45.828 INFO:teuthology.orchestra.run.vm03.stdout:alertmanager.vm03 vm03 *:9093,9094 running (2m) 5s ago 3m 14.7M - 0.25.0 c8568f914cd2 dfe36982fd82 2026-03-06T22:19:45.828 INFO:teuthology.orchestra.run.vm03.stdout:ceph-exporter.vm03 vm03 running (3m) 5s ago 3m 8836k - 19.2.3-39-g340d3c24fc6 8bccc98d839a f09a38798cc8 2026-03-06T22:19:45.828 INFO:teuthology.orchestra.run.vm03.stdout:ceph-exporter.vm08 vm08 running (2m) 4s ago 2m 6068k - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53153dc88f5a 2026-03-06T22:19:45.828 INFO:teuthology.orchestra.run.vm03.stdout:crash.vm03 vm03 running (3m) 5s ago 3m 12.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 708f937ce517 2026-03-06T22:19:45.828 INFO:teuthology.orchestra.run.vm03.stdout:crash.vm08 vm08 running (2m) 4s ago 2m 10.7M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 2bb962da32ac 2026-03-06T22:19:45.828 INFO:teuthology.orchestra.run.vm03.stdout:grafana.vm03 vm03 *:3000 running (2m) 5s ago 2m 64.1M - 10.4.0 c8b91775d855 6bf553fcdf8b 2026-03-06T22:19:45.828 INFO:teuthology.orchestra.run.vm03.stdout:mgr.vm03.uwuzgl vm03 *:9283,8765,8443 running (4m) 5s ago 4m 527M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 446d797c0836 2026-03-06T22:19:45.828 INFO:teuthology.orchestra.run.vm03.stdout:mgr.vm08.tdhano vm08 *:8443,9283,8765 running (2m) 4s ago 2m 472M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 18f9a0fe4daa 2026-03-06T22:19:45.828 INFO:teuthology.orchestra.run.vm03.stdout:mon.vm03 vm03 running (4m) 5s ago 4m 47.7M 2048M 19.2.3-39-g340d3c24fc6 8bccc98d839a dedab1238dc2 2026-03-06T22:19:45.828 INFO:teuthology.orchestra.run.vm03.stdout:mon.vm08 vm08 running (2m) 4s ago 2m 41.0M 2048M 19.2.3-39-g340d3c24fc6 8bccc98d839a e8de0b813933 2026-03-06T22:19:45.828 INFO:teuthology.orchestra.run.vm03.stdout:node-exporter.vm03 vm03 *:9100 running (3m) 5s ago 3m 7388k - 1.7.0 72c9c2088986 d631da09166b 2026-03-06T22:19:45.828 INFO:teuthology.orchestra.run.vm03.stdout:node-exporter.vm08 vm08 *:9100 running (2m) 4s ago 2m 7604k - 1.7.0 72c9c2088986 678b1f270c8e 2026-03-06T22:19:45.829 INFO:teuthology.orchestra.run.vm03.stdout:osd.0 vm08 running (93s) 4s ago 96s 34.7M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 5f735e647563 2026-03-06T22:19:45.829 INFO:teuthology.orchestra.run.vm03.stdout:osd.1 vm03 running (93s) 5s ago 96s 37.2M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 82414a348fdc 2026-03-06T22:19:45.829 INFO:teuthology.orchestra.run.vm03.stdout:osd.2 vm08 running (91s) 4s ago 94s 36.8M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a a22a71708d83 2026-03-06T22:19:45.829 INFO:teuthology.orchestra.run.vm03.stdout:osd.3 vm03 running (90s) 5s ago 94s 58.8M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a f968e8041eb8 2026-03-06T22:19:45.829 INFO:teuthology.orchestra.run.vm03.stdout:osd.4 vm08 running (89s) 4s ago 92s 34.7M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a e1c933a67e93 2026-03-06T22:19:45.829 INFO:teuthology.orchestra.run.vm03.stdout:osd.5 vm03 running (88s) 5s ago 92s 37.4M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a b50ce1f74566 2026-03-06T22:19:45.829 INFO:teuthology.orchestra.run.vm03.stdout:osd.6 vm08 running (87s) 4s ago 90s 36.2M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 7a747b887e5b 2026-03-06T22:19:45.829 INFO:teuthology.orchestra.run.vm03.stdout:osd.7 vm03 running (87s) 5s ago 90s 58.2M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 1e529bba7060 2026-03-06T22:19:45.829 INFO:teuthology.orchestra.run.vm03.stdout:prometheus.vm03 vm03 *:9095 running (2m) 5s ago 2m 35.2M - 2.51.0 1d3b7f56885b 8c9d328256b0 2026-03-06T22:19:45.897 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- bash -c 'ceph orch ls' 2026-03-06T22:19:47.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:47 vm03 bash[17055]: audit 2026-03-06T21:19:45.818679+0000 mgr.vm03.uwuzgl (mgr.14199) 140 : audit [DBG] from='client.14450 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:19:47.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:47 vm03 bash[17055]: audit 2026-03-06T21:19:45.818679+0000 mgr.vm03.uwuzgl (mgr.14199) 140 : audit [DBG] from='client.14450 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:19:47.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:47 vm03 bash[17055]: cluster 2026-03-06T21:19:46.081561+0000 mgr.vm03.uwuzgl (mgr.14199) 141 : cluster [DBG] pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:47.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:47 vm03 bash[17055]: cluster 2026-03-06T21:19:46.081561+0000 mgr.vm03.uwuzgl (mgr.14199) 141 : cluster [DBG] pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:47.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:47 vm08 bash[20828]: audit 2026-03-06T21:19:45.818679+0000 mgr.vm03.uwuzgl (mgr.14199) 140 : audit [DBG] from='client.14450 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:19:47.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:47 vm08 bash[20828]: audit 2026-03-06T21:19:45.818679+0000 mgr.vm03.uwuzgl (mgr.14199) 140 : audit [DBG] from='client.14450 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:19:47.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:47 vm08 bash[20828]: cluster 2026-03-06T21:19:46.081561+0000 mgr.vm03.uwuzgl (mgr.14199) 141 : cluster [DBG] pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:47.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:47 vm08 bash[20828]: cluster 2026-03-06T21:19:46.081561+0000 mgr.vm03.uwuzgl (mgr.14199) 141 : cluster [DBG] pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:49.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:49 vm08 bash[20828]: cluster 2026-03-06T21:19:48.081826+0000 mgr.vm03.uwuzgl (mgr.14199) 142 : cluster [DBG] pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:49.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:49 vm08 bash[20828]: cluster 2026-03-06T21:19:48.081826+0000 mgr.vm03.uwuzgl (mgr.14199) 142 : cluster [DBG] pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:49.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:49 vm03 bash[17055]: cluster 2026-03-06T21:19:48.081826+0000 mgr.vm03.uwuzgl (mgr.14199) 142 : cluster [DBG] pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:49.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:49 vm03 bash[17055]: cluster 2026-03-06T21:19:48.081826+0000 mgr.vm03.uwuzgl (mgr.14199) 142 : cluster [DBG] pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:50.695 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:51.057 INFO:teuthology.orchestra.run.vm03.stdout:NAME PORTS RUNNING REFRESHED AGE PLACEMENT 2026-03-06T22:19:51.057 INFO:teuthology.orchestra.run.vm03.stdout:alertmanager ?:9093,9094 1/1 10s ago 3m count:1 2026-03-06T22:19:51.057 INFO:teuthology.orchestra.run.vm03.stdout:ceph-exporter 2/2 10s ago 3m * 2026-03-06T22:19:51.057 INFO:teuthology.orchestra.run.vm03.stdout:crash 2/2 10s ago 3m * 2026-03-06T22:19:51.057 INFO:teuthology.orchestra.run.vm03.stdout:grafana ?:3000 1/1 10s ago 3m count:1 2026-03-06T22:19:51.057 INFO:teuthology.orchestra.run.vm03.stdout:mgr 2/2 10s ago 3m count:2 2026-03-06T22:19:51.057 INFO:teuthology.orchestra.run.vm03.stdout:mon 2/2 10s ago 2m vm03:192.168.123.103=vm03;vm08:192.168.123.108=vm08;count:2 2026-03-06T22:19:51.057 INFO:teuthology.orchestra.run.vm03.stdout:node-exporter ?:9100 2/2 10s ago 3m * 2026-03-06T22:19:51.057 INFO:teuthology.orchestra.run.vm03.stdout:osd.all-available-devices 8 10s ago 2m * 2026-03-06T22:19:51.057 INFO:teuthology.orchestra.run.vm03.stdout:prometheus ?:9095 1/1 10s ago 3m count:1 2026-03-06T22:19:51.129 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- bash -c 'ceph orch host ls' 2026-03-06T22:19:51.706 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:51 vm03 bash[17055]: cluster 2026-03-06T21:19:50.082051+0000 mgr.vm03.uwuzgl (mgr.14199) 143 : cluster [DBG] pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:51.706 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:51 vm03 bash[17055]: cluster 2026-03-06T21:19:50.082051+0000 mgr.vm03.uwuzgl (mgr.14199) 143 : cluster [DBG] pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:51.706 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:51 vm03 bash[17055]: audit 2026-03-06T21:19:51.058193+0000 mon.vm03 (mon.0) 631 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:19:51.706 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:51 vm03 bash[17055]: audit 2026-03-06T21:19:51.058193+0000 mon.vm03 (mon.0) 631 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:19:51.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:51 vm08 bash[20828]: cluster 2026-03-06T21:19:50.082051+0000 mgr.vm03.uwuzgl (mgr.14199) 143 : cluster [DBG] pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:51.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:51 vm08 bash[20828]: cluster 2026-03-06T21:19:50.082051+0000 mgr.vm03.uwuzgl (mgr.14199) 143 : cluster [DBG] pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:51.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:51 vm08 bash[20828]: audit 2026-03-06T21:19:51.058193+0000 mon.vm03 (mon.0) 631 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:19:51.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:51 vm08 bash[20828]: audit 2026-03-06T21:19:51.058193+0000 mon.vm03 (mon.0) 631 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:19:52.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:52 vm03 bash[17055]: audit 2026-03-06T21:19:51.049326+0000 mgr.vm03.uwuzgl (mgr.14199) 144 : audit [DBG] from='client.14454 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:19:52.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:52 vm03 bash[17055]: audit 2026-03-06T21:19:51.049326+0000 mgr.vm03.uwuzgl (mgr.14199) 144 : audit [DBG] from='client.14454 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:19:52.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:52 vm08 bash[20828]: audit 2026-03-06T21:19:51.049326+0000 mgr.vm03.uwuzgl (mgr.14199) 144 : audit [DBG] from='client.14454 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:19:52.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:52 vm08 bash[20828]: audit 2026-03-06T21:19:51.049326+0000 mgr.vm03.uwuzgl (mgr.14199) 144 : audit [DBG] from='client.14454 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:19:53.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:53 vm08 bash[20828]: cluster 2026-03-06T21:19:52.082298+0000 mgr.vm03.uwuzgl (mgr.14199) 145 : cluster [DBG] pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:53.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:53 vm08 bash[20828]: cluster 2026-03-06T21:19:52.082298+0000 mgr.vm03.uwuzgl (mgr.14199) 145 : cluster [DBG] pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:53.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:53 vm03 bash[17055]: cluster 2026-03-06T21:19:52.082298+0000 mgr.vm03.uwuzgl (mgr.14199) 145 : cluster [DBG] pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:53.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:53 vm03 bash[17055]: cluster 2026-03-06T21:19:52.082298+0000 mgr.vm03.uwuzgl (mgr.14199) 145 : cluster [DBG] pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:55.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:55 vm08 bash[20828]: cluster 2026-03-06T21:19:54.082584+0000 mgr.vm03.uwuzgl (mgr.14199) 146 : cluster [DBG] pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:55.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:55 vm08 bash[20828]: cluster 2026-03-06T21:19:54.082584+0000 mgr.vm03.uwuzgl (mgr.14199) 146 : cluster [DBG] pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:55.905 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:19:55.921 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:55 vm03 bash[17055]: cluster 2026-03-06T21:19:54.082584+0000 mgr.vm03.uwuzgl (mgr.14199) 146 : cluster [DBG] pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:55.921 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:55 vm03 bash[17055]: cluster 2026-03-06T21:19:54.082584+0000 mgr.vm03.uwuzgl (mgr.14199) 146 : cluster [DBG] pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:56.293 INFO:teuthology.orchestra.run.vm03.stdout:HOST ADDR LABELS STATUS 2026-03-06T22:19:56.293 INFO:teuthology.orchestra.run.vm03.stdout:vm03 192.168.123.103 2026-03-06T22:19:56.293 INFO:teuthology.orchestra.run.vm03.stdout:vm08 192.168.123.108 2026-03-06T22:19:56.293 INFO:teuthology.orchestra.run.vm03.stdout:2 hosts in cluster 2026-03-06T22:19:56.362 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- bash -c 'ceph orch device ls' 2026-03-06T22:19:57.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:57 vm08 bash[20828]: cluster 2026-03-06T21:19:56.082852+0000 mgr.vm03.uwuzgl (mgr.14199) 147 : cluster [DBG] pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:57.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:57 vm08 bash[20828]: cluster 2026-03-06T21:19:56.082852+0000 mgr.vm03.uwuzgl (mgr.14199) 147 : cluster [DBG] pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:57.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:57 vm08 bash[20828]: audit 2026-03-06T21:19:56.287261+0000 mgr.vm03.uwuzgl (mgr.14199) 148 : audit [DBG] from='client.14458 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:19:57.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:57 vm08 bash[20828]: audit 2026-03-06T21:19:56.287261+0000 mgr.vm03.uwuzgl (mgr.14199) 148 : audit [DBG] from='client.14458 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:19:57.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:57 vm03 bash[17055]: cluster 2026-03-06T21:19:56.082852+0000 mgr.vm03.uwuzgl (mgr.14199) 147 : cluster [DBG] pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:57.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:57 vm03 bash[17055]: cluster 2026-03-06T21:19:56.082852+0000 mgr.vm03.uwuzgl (mgr.14199) 147 : cluster [DBG] pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:57.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:57 vm03 bash[17055]: audit 2026-03-06T21:19:56.287261+0000 mgr.vm03.uwuzgl (mgr.14199) 148 : audit [DBG] from='client.14458 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:19:57.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:57 vm03 bash[17055]: audit 2026-03-06T21:19:56.287261+0000 mgr.vm03.uwuzgl (mgr.14199) 148 : audit [DBG] from='client.14458 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:19:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:59 vm08 bash[20828]: cluster 2026-03-06T21:19:58.083132+0000 mgr.vm03.uwuzgl (mgr.14199) 149 : cluster [DBG] pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:19:59 vm08 bash[20828]: cluster 2026-03-06T21:19:58.083132+0000 mgr.vm03.uwuzgl (mgr.14199) 149 : cluster [DBG] pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:59.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:59 vm03 bash[17055]: cluster 2026-03-06T21:19:58.083132+0000 mgr.vm03.uwuzgl (mgr.14199) 149 : cluster [DBG] pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:19:59.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:19:59 vm03 bash[17055]: cluster 2026-03-06T21:19:58.083132+0000 mgr.vm03.uwuzgl (mgr.14199) 149 : cluster [DBG] pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:00.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:00 vm08 bash[20828]: cluster 2026-03-06T21:20:00.000122+0000 mon.vm03 (mon.0) 632 : cluster [INF] overall HEALTH_OK 2026-03-06T22:20:00.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:00 vm08 bash[20828]: cluster 2026-03-06T21:20:00.000122+0000 mon.vm03 (mon.0) 632 : cluster [INF] overall HEALTH_OK 2026-03-06T22:20:00.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:00 vm03 bash[17055]: cluster 2026-03-06T21:20:00.000122+0000 mon.vm03 (mon.0) 632 : cluster [INF] overall HEALTH_OK 2026-03-06T22:20:00.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:00 vm03 bash[17055]: cluster 2026-03-06T21:20:00.000122+0000 mon.vm03 (mon.0) 632 : cluster [INF] overall HEALTH_OK 2026-03-06T22:20:01.163 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:20:01.546 INFO:teuthology.orchestra.run.vm03.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REFRESHED REJECT REASONS 2026-03-06T22:20:01.546 INFO:teuthology.orchestra.run.vm03.stdout:vm03 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 85s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-06T22:20:01.546 INFO:teuthology.orchestra.run.vm03.stdout:vm03 /dev/vdb hdd DWNBRSTVMM03001 20.0G No 85s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:20:01.546 INFO:teuthology.orchestra.run.vm03.stdout:vm03 /dev/vdc hdd DWNBRSTVMM03002 20.0G No 85s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:20:01.547 INFO:teuthology.orchestra.run.vm03.stdout:vm03 /dev/vdd hdd DWNBRSTVMM03003 20.0G No 85s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:20:01.547 INFO:teuthology.orchestra.run.vm03.stdout:vm03 /dev/vde hdd DWNBRSTVMM03004 20.0G No 85s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:20:01.547 INFO:teuthology.orchestra.run.vm03.stdout:vm08 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 85s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-06T22:20:01.547 INFO:teuthology.orchestra.run.vm03.stdout:vm08 /dev/vdb hdd DWNBRSTVMM08001 20.0G No 85s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:20:01.547 INFO:teuthology.orchestra.run.vm03.stdout:vm08 /dev/vdc hdd DWNBRSTVMM08002 20.0G No 85s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:20:01.547 INFO:teuthology.orchestra.run.vm03.stdout:vm08 /dev/vdd hdd DWNBRSTVMM08003 20.0G No 85s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:20:01.547 INFO:teuthology.orchestra.run.vm03.stdout:vm08 /dev/vde hdd DWNBRSTVMM08004 20.0G No 85s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:20:01.613 INFO:teuthology.run_tasks:Running task vip... 2026-03-06T22:20:01.616 INFO:tasks.vip:Allocating static IPs for each host... 2026-03-06T22:20:01.616 INFO:tasks.vip:peername 192.168.123.103 2026-03-06T22:20:01.616 INFO:tasks.vip:192.168.123.103 in 192.168.123.0/24, pos 102 2026-03-06T22:20:01.616 INFO:tasks.vip:vm03.local static 12.12.0.103, vnet 12.12.0.0/22 2026-03-06T22:20:01.616 INFO:tasks.vip:VIPs are [IPv4Address('12.12.1.103')] 2026-03-06T22:20:01.616 DEBUG:teuthology.orchestra.run.vm03:> sudo ip route ls 2026-03-06T22:20:01.624 INFO:teuthology.orchestra.run.vm03.stdout:default via 192.168.123.1 dev ens3 proto dhcp src 192.168.123.103 metric 100 2026-03-06T22:20:01.624 INFO:teuthology.orchestra.run.vm03.stdout:172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.0.1 linkdown 2026-03-06T22:20:01.624 INFO:teuthology.orchestra.run.vm03.stdout:192.168.123.0/24 dev ens3 proto kernel scope link src 192.168.123.103 metric 100 2026-03-06T22:20:01.624 INFO:teuthology.orchestra.run.vm03.stdout:192.168.123.1 dev ens3 proto dhcp scope link src 192.168.123.103 metric 100 2026-03-06T22:20:01.624 INFO:tasks.vip:Configuring 12.12.0.103 on vm03.local iface ens3... 2026-03-06T22:20:01.624 DEBUG:teuthology.orchestra.run.vm03:> sudo ip addr add 12.12.0.103/22 dev ens3 2026-03-06T22:20:01.673 INFO:tasks.vip:peername 192.168.123.108 2026-03-06T22:20:01.673 INFO:tasks.vip:192.168.123.108 in 192.168.123.0/24, pos 107 2026-03-06T22:20:01.673 INFO:tasks.vip:vm08.local static 12.12.0.108, vnet 12.12.0.0/22 2026-03-06T22:20:01.673 DEBUG:teuthology.orchestra.run.vm08:> sudo ip route ls 2026-03-06T22:20:01.680 INFO:teuthology.orchestra.run.vm08.stdout:default via 192.168.123.1 dev ens3 proto dhcp src 192.168.123.108 metric 100 2026-03-06T22:20:01.680 INFO:teuthology.orchestra.run.vm08.stdout:172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.0.1 linkdown 2026-03-06T22:20:01.680 INFO:teuthology.orchestra.run.vm08.stdout:192.168.123.0/24 dev ens3 proto kernel scope link src 192.168.123.108 metric 100 2026-03-06T22:20:01.680 INFO:teuthology.orchestra.run.vm08.stdout:192.168.123.1 dev ens3 proto dhcp scope link src 192.168.123.108 metric 100 2026-03-06T22:20:01.680 INFO:tasks.vip:Configuring 12.12.0.108 on vm08.local iface ens3... 2026-03-06T22:20:01.680 DEBUG:teuthology.orchestra.run.vm08:> sudo ip addr add 12.12.0.108/22 dev ens3 2026-03-06T22:20:01.730 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-06T22:20:01.732 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm03.local 2026-03-06T22:20:01.732 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- bash -c 'ceph orch device ls --refresh' 2026-03-06T22:20:01.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:01 vm08 bash[20828]: cluster 2026-03-06T21:20:00.083395+0000 mgr.vm03.uwuzgl (mgr.14199) 150 : cluster [DBG] pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:01.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:01 vm08 bash[20828]: cluster 2026-03-06T21:20:00.083395+0000 mgr.vm03.uwuzgl (mgr.14199) 150 : cluster [DBG] pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:01.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:01 vm03 bash[17055]: cluster 2026-03-06T21:20:00.083395+0000 mgr.vm03.uwuzgl (mgr.14199) 150 : cluster [DBG] pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:01.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:01 vm03 bash[17055]: cluster 2026-03-06T21:20:00.083395+0000 mgr.vm03.uwuzgl (mgr.14199) 150 : cluster [DBG] pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:02.871 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:02 vm03 bash[17055]: audit 2026-03-06T21:20:01.539863+0000 mgr.vm03.uwuzgl (mgr.14199) 151 : audit [DBG] from='client.24301 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:02.872 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:02 vm03 bash[17055]: audit 2026-03-06T21:20:01.539863+0000 mgr.vm03.uwuzgl (mgr.14199) 151 : audit [DBG] from='client.24301 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:02.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:02 vm08 bash[20828]: audit 2026-03-06T21:20:01.539863+0000 mgr.vm03.uwuzgl (mgr.14199) 151 : audit [DBG] from='client.24301 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:02.900 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:02 vm08 bash[20828]: audit 2026-03-06T21:20:01.539863+0000 mgr.vm03.uwuzgl (mgr.14199) 151 : audit [DBG] from='client.24301 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:03.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:03 vm03 bash[17055]: cluster 2026-03-06T21:20:02.083699+0000 mgr.vm03.uwuzgl (mgr.14199) 152 : cluster [DBG] pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:03.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:03 vm03 bash[17055]: cluster 2026-03-06T21:20:02.083699+0000 mgr.vm03.uwuzgl (mgr.14199) 152 : cluster [DBG] pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:04.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:03 vm08 bash[20828]: cluster 2026-03-06T21:20:02.083699+0000 mgr.vm03.uwuzgl (mgr.14199) 152 : cluster [DBG] pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:04.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:03 vm08 bash[20828]: cluster 2026-03-06T21:20:02.083699+0000 mgr.vm03.uwuzgl (mgr.14199) 152 : cluster [DBG] pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:05.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:05 vm03 bash[17055]: cluster 2026-03-06T21:20:04.083993+0000 mgr.vm03.uwuzgl (mgr.14199) 153 : cluster [DBG] pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:05.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:05 vm03 bash[17055]: cluster 2026-03-06T21:20:04.083993+0000 mgr.vm03.uwuzgl (mgr.14199) 153 : cluster [DBG] pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:06.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:05 vm08 bash[20828]: cluster 2026-03-06T21:20:04.083993+0000 mgr.vm03.uwuzgl (mgr.14199) 153 : cluster [DBG] pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:06.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:05 vm08 bash[20828]: cluster 2026-03-06T21:20:04.083993+0000 mgr.vm03.uwuzgl (mgr.14199) 153 : cluster [DBG] pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:06.507 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:20:06.871 INFO:teuthology.orchestra.run.vm03.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REFRESHED REJECT REASONS 2026-03-06T22:20:06.871 INFO:teuthology.orchestra.run.vm03.stdout:vm03 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 91s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-06T22:20:06.871 INFO:teuthology.orchestra.run.vm03.stdout:vm03 /dev/vdb hdd DWNBRSTVMM03001 20.0G No 91s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:20:06.871 INFO:teuthology.orchestra.run.vm03.stdout:vm03 /dev/vdc hdd DWNBRSTVMM03002 20.0G No 91s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:20:06.871 INFO:teuthology.orchestra.run.vm03.stdout:vm03 /dev/vdd hdd DWNBRSTVMM03003 20.0G No 91s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:20:06.871 INFO:teuthology.orchestra.run.vm03.stdout:vm03 /dev/vde hdd DWNBRSTVMM03004 20.0G No 91s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:20:06.871 INFO:teuthology.orchestra.run.vm03.stdout:vm08 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 91s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-06T22:20:06.871 INFO:teuthology.orchestra.run.vm03.stdout:vm08 /dev/vdb hdd DWNBRSTVMM08001 20.0G No 91s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:20:06.871 INFO:teuthology.orchestra.run.vm03.stdout:vm08 /dev/vdc hdd DWNBRSTVMM08002 20.0G No 91s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:20:06.871 INFO:teuthology.orchestra.run.vm03.stdout:vm08 /dev/vdd hdd DWNBRSTVMM08003 20.0G No 91s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:20:06.872 INFO:teuthology.orchestra.run.vm03.stdout:vm08 /dev/vde hdd DWNBRSTVMM08004 20.0G No 91s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:20:06.903 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:06 vm03 bash[17055]: audit 2026-03-06T21:20:06.058361+0000 mon.vm03 (mon.0) 633 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:20:06.903 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:06 vm03 bash[17055]: audit 2026-03-06T21:20:06.058361+0000 mon.vm03 (mon.0) 633 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:20:06.964 INFO:teuthology.run_tasks:Running task vip.exec... 2026-03-06T22:20:06.966 INFO:tasks.vip:Running commands on role host.a host ubuntu@vm03.local 2026-03-06T22:20:06.966 DEBUG:teuthology.orchestra.run.vm03:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'systemctl stop nfs-server' 2026-03-06T22:20:06.973 INFO:teuthology.orchestra.run.vm03.stderr:+ systemctl stop nfs-server 2026-03-06T22:20:06.976 INFO:tasks.vip:Running commands on role host.b host ubuntu@vm08.local 2026-03-06T22:20:06.976 DEBUG:teuthology.orchestra.run.vm08:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'systemctl stop nfs-server' 2026-03-06T22:20:06.983 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:06 vm08 bash[20828]: audit 2026-03-06T21:20:06.058361+0000 mon.vm03 (mon.0) 633 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:20:06.983 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:06 vm08 bash[20828]: audit 2026-03-06T21:20:06.058361+0000 mon.vm03 (mon.0) 633 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:20:06.985 INFO:teuthology.orchestra.run.vm08.stderr:+ systemctl stop nfs-server 2026-03-06T22:20:06.989 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-06T22:20:06.992 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm03.local 2026-03-06T22:20:06.992 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- bash -c 'ceph fs volume create fs1' 2026-03-06T22:20:07.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:07 vm03 bash[17055]: cluster 2026-03-06T21:20:06.084233+0000 mgr.vm03.uwuzgl (mgr.14199) 154 : cluster [DBG] pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:07.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:07 vm03 bash[17055]: cluster 2026-03-06T21:20:06.084233+0000 mgr.vm03.uwuzgl (mgr.14199) 154 : cluster [DBG] pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:07.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:07 vm03 bash[17055]: audit 2026-03-06T21:20:06.865126+0000 mon.vm03 (mon.0) 634 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:20:07.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:07 vm03 bash[17055]: audit 2026-03-06T21:20:06.865126+0000 mon.vm03 (mon.0) 634 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:20:08.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:07 vm08 bash[20828]: cluster 2026-03-06T21:20:06.084233+0000 mgr.vm03.uwuzgl (mgr.14199) 154 : cluster [DBG] pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:08.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:07 vm08 bash[20828]: cluster 2026-03-06T21:20:06.084233+0000 mgr.vm03.uwuzgl (mgr.14199) 154 : cluster [DBG] pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:08.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:07 vm08 bash[20828]: audit 2026-03-06T21:20:06.865126+0000 mon.vm03 (mon.0) 634 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:20:08.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:07 vm08 bash[20828]: audit 2026-03-06T21:20:06.865126+0000 mon.vm03 (mon.0) 634 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:20:08.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:08 vm03 bash[17055]: audit 2026-03-06T21:20:06.863468+0000 mgr.vm03.uwuzgl (mgr.14199) 155 : audit [DBG] from='client.14466 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:08.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:08 vm03 bash[17055]: audit 2026-03-06T21:20:06.863468+0000 mgr.vm03.uwuzgl (mgr.14199) 155 : audit [DBG] from='client.14466 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:09.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:08 vm08 bash[20828]: audit 2026-03-06T21:20:06.863468+0000 mgr.vm03.uwuzgl (mgr.14199) 155 : audit [DBG] from='client.14466 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:09.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:08 vm08 bash[20828]: audit 2026-03-06T21:20:06.863468+0000 mgr.vm03.uwuzgl (mgr.14199) 155 : audit [DBG] from='client.14466 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:09.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:09 vm03 bash[17055]: cluster 2026-03-06T21:20:08.084461+0000 mgr.vm03.uwuzgl (mgr.14199) 156 : cluster [DBG] pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:09.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:09 vm03 bash[17055]: cluster 2026-03-06T21:20:08.084461+0000 mgr.vm03.uwuzgl (mgr.14199) 156 : cluster [DBG] pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:10.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:09 vm08 bash[20828]: cluster 2026-03-06T21:20:08.084461+0000 mgr.vm03.uwuzgl (mgr.14199) 156 : cluster [DBG] pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:10.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:09 vm08 bash[20828]: cluster 2026-03-06T21:20:08.084461+0000 mgr.vm03.uwuzgl (mgr.14199) 156 : cluster [DBG] pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:10.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:10 vm03 bash[17055]: cluster 2026-03-06T21:20:10.084683+0000 mgr.vm03.uwuzgl (mgr.14199) 157 : cluster [DBG] pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:10.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:10 vm03 bash[17055]: cluster 2026-03-06T21:20:10.084683+0000 mgr.vm03.uwuzgl (mgr.14199) 157 : cluster [DBG] pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:11.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:10 vm08 bash[20828]: cluster 2026-03-06T21:20:10.084683+0000 mgr.vm03.uwuzgl (mgr.14199) 157 : cluster [DBG] pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:11.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:10 vm08 bash[20828]: cluster 2026-03-06T21:20:10.084683+0000 mgr.vm03.uwuzgl (mgr.14199) 157 : cluster [DBG] pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:11.709 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:20:13.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:11.965935+0000 mon.vm03 (mon.0) 635 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:11.965935+0000 mon.vm03 (mon.0) 635 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:11.970698+0000 mon.vm03 (mon.0) 636 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:11.970698+0000 mon.vm03 (mon.0) 636 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: cluster 2026-03-06T21:20:12.084993+0000 mgr.vm03.uwuzgl (mgr.14199) 158 : cluster [DBG] pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: cluster 2026-03-06T21:20:12.084993+0000 mgr.vm03.uwuzgl (mgr.14199) 158 : cluster [DBG] pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:12.245638+0000 mgr.vm03.uwuzgl (mgr.14199) 159 : audit [DBG] from='client.14470 -' entity='client.admin' cmd=[{"prefix": "fs volume create", "name": "fs1", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:12.245638+0000 mgr.vm03.uwuzgl (mgr.14199) 159 : audit [DBG] from='client.14470 -' entity='client.admin' cmd=[{"prefix": "fs volume create", "name": "fs1", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:12.246246+0000 mon.vm03 (mon.0) 637 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool create", "pool": "cephfs.fs1.meta"}]: dispatch 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:12.246246+0000 mon.vm03 (mon.0) 637 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool create", "pool": "cephfs.fs1.meta"}]: dispatch 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:12.269910+0000 mon.vm03 (mon.0) 638 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:12.269910+0000 mon.vm03 (mon.0) 638 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:12.274773+0000 mon.vm03 (mon.0) 639 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:12.274773+0000 mon.vm03 (mon.0) 639 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:12.459464+0000 mon.vm03 (mon.0) 640 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:12.459464+0000 mon.vm03 (mon.0) 640 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:12.465741+0000 mon.vm03 (mon.0) 641 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:12.465741+0000 mon.vm03 (mon.0) 641 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:12.736788+0000 mon.vm03 (mon.0) 642 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:12.736788+0000 mon.vm03 (mon.0) 642 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:12.742639+0000 mon.vm03 (mon.0) 643 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:12 vm08 bash[20828]: audit 2026-03-06T21:20:12.742639+0000 mon.vm03 (mon.0) 643 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:11.965935+0000 mon.vm03 (mon.0) 635 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:11.965935+0000 mon.vm03 (mon.0) 635 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:11.970698+0000 mon.vm03 (mon.0) 636 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:11.970698+0000 mon.vm03 (mon.0) 636 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: cluster 2026-03-06T21:20:12.084993+0000 mgr.vm03.uwuzgl (mgr.14199) 158 : cluster [DBG] pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: cluster 2026-03-06T21:20:12.084993+0000 mgr.vm03.uwuzgl (mgr.14199) 158 : cluster [DBG] pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:12.245638+0000 mgr.vm03.uwuzgl (mgr.14199) 159 : audit [DBG] from='client.14470 -' entity='client.admin' cmd=[{"prefix": "fs volume create", "name": "fs1", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:12.245638+0000 mgr.vm03.uwuzgl (mgr.14199) 159 : audit [DBG] from='client.14470 -' entity='client.admin' cmd=[{"prefix": "fs volume create", "name": "fs1", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:12.246246+0000 mon.vm03 (mon.0) 637 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool create", "pool": "cephfs.fs1.meta"}]: dispatch 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:12.246246+0000 mon.vm03 (mon.0) 637 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool create", "pool": "cephfs.fs1.meta"}]: dispatch 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:12.269910+0000 mon.vm03 (mon.0) 638 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:12.269910+0000 mon.vm03 (mon.0) 638 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:12.274773+0000 mon.vm03 (mon.0) 639 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:12.274773+0000 mon.vm03 (mon.0) 639 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:12.459464+0000 mon.vm03 (mon.0) 640 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:12.459464+0000 mon.vm03 (mon.0) 640 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:12.465741+0000 mon.vm03 (mon.0) 641 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:12.465741+0000 mon.vm03 (mon.0) 641 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:12.736788+0000 mon.vm03 (mon.0) 642 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:12.736788+0000 mon.vm03 (mon.0) 642 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:12.742639+0000 mon.vm03 (mon.0) 643 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:13.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:12 vm03 bash[17055]: audit 2026-03-06T21:20:12.742639+0000 mon.vm03 (mon.0) 643 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:14.157 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- bash -c 'ceph nfs cluster create happy --ingress --virtual-ip=12.12.1.103 --ingress-mode=haproxy-protocol' 2026-03-06T22:20:14.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:14 vm03 bash[17055]: debug 2026-03-06T21:20:14.012+0000 7ff816041640 -1 log_channel(cluster) log [ERR] : Health check failed: 1 filesystem is offline (MDS_ALL_DOWN) 2026-03-06T22:20:14.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:14 vm03 bash[17055]: audit 2026-03-06T21:20:12.983454+0000 mon.vm03 (mon.0) 644 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool create", "pool": "cephfs.fs1.meta"}]': finished 2026-03-06T22:20:14.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:14 vm03 bash[17055]: audit 2026-03-06T21:20:12.983454+0000 mon.vm03 (mon.0) 644 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool create", "pool": "cephfs.fs1.meta"}]': finished 2026-03-06T22:20:14.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:14 vm03 bash[17055]: cluster 2026-03-06T21:20:12.991439+0000 mon.vm03 (mon.0) 645 : cluster [DBG] osdmap e25: 8 total, 8 up, 8 in 2026-03-06T22:20:14.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:14 vm03 bash[17055]: cluster 2026-03-06T21:20:12.991439+0000 mon.vm03 (mon.0) 645 : cluster [DBG] osdmap e25: 8 total, 8 up, 8 in 2026-03-06T22:20:14.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:14 vm03 bash[17055]: audit 2026-03-06T21:20:13.009675+0000 mon.vm03 (mon.0) 646 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"bulk": true, "prefix": "osd pool create", "pool": "cephfs.fs1.data"}]: dispatch 2026-03-06T22:20:14.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:14 vm03 bash[17055]: audit 2026-03-06T21:20:13.009675+0000 mon.vm03 (mon.0) 646 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"bulk": true, "prefix": "osd pool create", "pool": "cephfs.fs1.data"}]: dispatch 2026-03-06T22:20:14.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:14 vm08 bash[20828]: audit 2026-03-06T21:20:12.983454+0000 mon.vm03 (mon.0) 644 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool create", "pool": "cephfs.fs1.meta"}]': finished 2026-03-06T22:20:14.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:14 vm08 bash[20828]: audit 2026-03-06T21:20:12.983454+0000 mon.vm03 (mon.0) 644 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool create", "pool": "cephfs.fs1.meta"}]': finished 2026-03-06T22:20:14.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:14 vm08 bash[20828]: cluster 2026-03-06T21:20:12.991439+0000 mon.vm03 (mon.0) 645 : cluster [DBG] osdmap e25: 8 total, 8 up, 8 in 2026-03-06T22:20:14.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:14 vm08 bash[20828]: cluster 2026-03-06T21:20:12.991439+0000 mon.vm03 (mon.0) 645 : cluster [DBG] osdmap e25: 8 total, 8 up, 8 in 2026-03-06T22:20:14.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:14 vm08 bash[20828]: audit 2026-03-06T21:20:13.009675+0000 mon.vm03 (mon.0) 646 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"bulk": true, "prefix": "osd pool create", "pool": "cephfs.fs1.data"}]: dispatch 2026-03-06T22:20:14.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:14 vm08 bash[20828]: audit 2026-03-06T21:20:13.009675+0000 mon.vm03 (mon.0) 646 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"bulk": true, "prefix": "osd pool create", "pool": "cephfs.fs1.data"}]: dispatch 2026-03-06T22:20:15.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: audit 2026-03-06T21:20:13.987973+0000 mon.vm03 (mon.0) 647 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"bulk": true, "prefix": "osd pool create", "pool": "cephfs.fs1.data"}]': finished 2026-03-06T22:20:15.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: audit 2026-03-06T21:20:13.987973+0000 mon.vm03 (mon.0) 647 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"bulk": true, "prefix": "osd pool create", "pool": "cephfs.fs1.data"}]': finished 2026-03-06T22:20:15.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: cluster 2026-03-06T21:20:13.990762+0000 mon.vm03 (mon.0) 648 : cluster [DBG] osdmap e26: 8 total, 8 up, 8 in 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: cluster 2026-03-06T21:20:13.990762+0000 mon.vm03 (mon.0) 648 : cluster [DBG] osdmap e26: 8 total, 8 up, 8 in 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: audit 2026-03-06T21:20:14.018335+0000 mon.vm03 (mon.0) 649 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "fs new", "fs_name": "fs1", "metadata": "cephfs.fs1.meta", "data": "cephfs.fs1.data"}]: dispatch 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: audit 2026-03-06T21:20:14.018335+0000 mon.vm03 (mon.0) 649 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "fs new", "fs_name": "fs1", "metadata": "cephfs.fs1.meta", "data": "cephfs.fs1.data"}]: dispatch 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: cluster 2026-03-06T21:20:14.018753+0000 mon.vm03 (mon.0) 650 : cluster [ERR] Health check failed: 1 filesystem is offline (MDS_ALL_DOWN) 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: cluster 2026-03-06T21:20:14.018753+0000 mon.vm03 (mon.0) 650 : cluster [ERR] Health check failed: 1 filesystem is offline (MDS_ALL_DOWN) 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: cluster 2026-03-06T21:20:14.018762+0000 mon.vm03 (mon.0) 651 : cluster [WRN] Health check failed: 1 filesystem is online with fewer MDS than max_mds (MDS_UP_LESS_THAN_MAX) 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: cluster 2026-03-06T21:20:14.018762+0000 mon.vm03 (mon.0) 651 : cluster [WRN] Health check failed: 1 filesystem is online with fewer MDS than max_mds (MDS_UP_LESS_THAN_MAX) 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: audit 2026-03-06T21:20:14.046576+0000 mon.vm03 (mon.0) 652 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "fs new", "fs_name": "fs1", "metadata": "cephfs.fs1.meta", "data": "cephfs.fs1.data"}]': finished 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: audit 2026-03-06T21:20:14.046576+0000 mon.vm03 (mon.0) 652 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "fs new", "fs_name": "fs1", "metadata": "cephfs.fs1.meta", "data": "cephfs.fs1.data"}]': finished 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: cluster 2026-03-06T21:20:14.062735+0000 mon.vm03 (mon.0) 653 : cluster [DBG] osdmap e27: 8 total, 8 up, 8 in 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: cluster 2026-03-06T21:20:14.062735+0000 mon.vm03 (mon.0) 653 : cluster [DBG] osdmap e27: 8 total, 8 up, 8 in 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: cluster 2026-03-06T21:20:14.063385+0000 mon.vm03 (mon.0) 654 : cluster [DBG] fsmap fs1:0 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: cluster 2026-03-06T21:20:14.063385+0000 mon.vm03 (mon.0) 654 : cluster [DBG] fsmap fs1:0 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: cephadm 2026-03-06T21:20:14.064873+0000 mgr.vm03.uwuzgl (mgr.14199) 160 : cephadm [INF] Saving service mds.fs1 spec with placement count:2 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: cephadm 2026-03-06T21:20:14.064873+0000 mgr.vm03.uwuzgl (mgr.14199) 160 : cephadm [INF] Saving service mds.fs1 spec with placement count:2 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: audit 2026-03-06T21:20:14.073759+0000 mon.vm03 (mon.0) 655 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: audit 2026-03-06T21:20:14.073759+0000 mon.vm03 (mon.0) 655 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: cluster 2026-03-06T21:20:14.085245+0000 mgr.vm03.uwuzgl (mgr.14199) 161 : cluster [DBG] pgmap v109: 65 pgs: 12 creating+peering, 1 active+clean, 52 unknown; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:15.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:15 vm03 bash[17055]: cluster 2026-03-06T21:20:14.085245+0000 mgr.vm03.uwuzgl (mgr.14199) 161 : cluster [DBG] pgmap v109: 65 pgs: 12 creating+peering, 1 active+clean, 52 unknown; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: audit 2026-03-06T21:20:13.987973+0000 mon.vm03 (mon.0) 647 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"bulk": true, "prefix": "osd pool create", "pool": "cephfs.fs1.data"}]': finished 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: audit 2026-03-06T21:20:13.987973+0000 mon.vm03 (mon.0) 647 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"bulk": true, "prefix": "osd pool create", "pool": "cephfs.fs1.data"}]': finished 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: cluster 2026-03-06T21:20:13.990762+0000 mon.vm03 (mon.0) 648 : cluster [DBG] osdmap e26: 8 total, 8 up, 8 in 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: cluster 2026-03-06T21:20:13.990762+0000 mon.vm03 (mon.0) 648 : cluster [DBG] osdmap e26: 8 total, 8 up, 8 in 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: audit 2026-03-06T21:20:14.018335+0000 mon.vm03 (mon.0) 649 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "fs new", "fs_name": "fs1", "metadata": "cephfs.fs1.meta", "data": "cephfs.fs1.data"}]: dispatch 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: audit 2026-03-06T21:20:14.018335+0000 mon.vm03 (mon.0) 649 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "fs new", "fs_name": "fs1", "metadata": "cephfs.fs1.meta", "data": "cephfs.fs1.data"}]: dispatch 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: cluster 2026-03-06T21:20:14.018753+0000 mon.vm03 (mon.0) 650 : cluster [ERR] Health check failed: 1 filesystem is offline (MDS_ALL_DOWN) 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: cluster 2026-03-06T21:20:14.018753+0000 mon.vm03 (mon.0) 650 : cluster [ERR] Health check failed: 1 filesystem is offline (MDS_ALL_DOWN) 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: cluster 2026-03-06T21:20:14.018762+0000 mon.vm03 (mon.0) 651 : cluster [WRN] Health check failed: 1 filesystem is online with fewer MDS than max_mds (MDS_UP_LESS_THAN_MAX) 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: cluster 2026-03-06T21:20:14.018762+0000 mon.vm03 (mon.0) 651 : cluster [WRN] Health check failed: 1 filesystem is online with fewer MDS than max_mds (MDS_UP_LESS_THAN_MAX) 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: audit 2026-03-06T21:20:14.046576+0000 mon.vm03 (mon.0) 652 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "fs new", "fs_name": "fs1", "metadata": "cephfs.fs1.meta", "data": "cephfs.fs1.data"}]': finished 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: audit 2026-03-06T21:20:14.046576+0000 mon.vm03 (mon.0) 652 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "fs new", "fs_name": "fs1", "metadata": "cephfs.fs1.meta", "data": "cephfs.fs1.data"}]': finished 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: cluster 2026-03-06T21:20:14.062735+0000 mon.vm03 (mon.0) 653 : cluster [DBG] osdmap e27: 8 total, 8 up, 8 in 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: cluster 2026-03-06T21:20:14.062735+0000 mon.vm03 (mon.0) 653 : cluster [DBG] osdmap e27: 8 total, 8 up, 8 in 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: cluster 2026-03-06T21:20:14.063385+0000 mon.vm03 (mon.0) 654 : cluster [DBG] fsmap fs1:0 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: cluster 2026-03-06T21:20:14.063385+0000 mon.vm03 (mon.0) 654 : cluster [DBG] fsmap fs1:0 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: cephadm 2026-03-06T21:20:14.064873+0000 mgr.vm03.uwuzgl (mgr.14199) 160 : cephadm [INF] Saving service mds.fs1 spec with placement count:2 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: cephadm 2026-03-06T21:20:14.064873+0000 mgr.vm03.uwuzgl (mgr.14199) 160 : cephadm [INF] Saving service mds.fs1 spec with placement count:2 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: audit 2026-03-06T21:20:14.073759+0000 mon.vm03 (mon.0) 655 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: audit 2026-03-06T21:20:14.073759+0000 mon.vm03 (mon.0) 655 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:15.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: cluster 2026-03-06T21:20:14.085245+0000 mgr.vm03.uwuzgl (mgr.14199) 161 : cluster [DBG] pgmap v109: 65 pgs: 12 creating+peering, 1 active+clean, 52 unknown; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:15.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:15 vm08 bash[20828]: cluster 2026-03-06T21:20:14.085245+0000 mgr.vm03.uwuzgl (mgr.14199) 161 : cluster [DBG] pgmap v109: 65 pgs: 12 creating+peering, 1 active+clean, 52 unknown; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:16.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:16 vm03 bash[17055]: cluster 2026-03-06T21:20:15.057107+0000 mon.vm03 (mon.0) 656 : cluster [DBG] osdmap e28: 8 total, 8 up, 8 in 2026-03-06T22:20:16.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:16 vm03 bash[17055]: cluster 2026-03-06T21:20:15.057107+0000 mon.vm03 (mon.0) 656 : cluster [DBG] osdmap e28: 8 total, 8 up, 8 in 2026-03-06T22:20:16.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:16 vm08 bash[20828]: cluster 2026-03-06T21:20:15.057107+0000 mon.vm03 (mon.0) 656 : cluster [DBG] osdmap e28: 8 total, 8 up, 8 in 2026-03-06T22:20:16.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:16 vm08 bash[20828]: cluster 2026-03-06T21:20:15.057107+0000 mon.vm03 (mon.0) 656 : cluster [DBG] osdmap e28: 8 total, 8 up, 8 in 2026-03-06T22:20:17.406 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:17 vm03 bash[17055]: cluster 2026-03-06T21:20:16.068519+0000 mon.vm03 (mon.0) 657 : cluster [DBG] osdmap e29: 8 total, 8 up, 8 in 2026-03-06T22:20:17.406 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:17 vm03 bash[17055]: cluster 2026-03-06T21:20:16.068519+0000 mon.vm03 (mon.0) 657 : cluster [DBG] osdmap e29: 8 total, 8 up, 8 in 2026-03-06T22:20:17.406 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:17 vm03 bash[17055]: cluster 2026-03-06T21:20:16.085563+0000 mgr.vm03.uwuzgl (mgr.14199) 162 : cluster [DBG] pgmap v112: 65 pgs: 5 creating+activating, 16 creating+peering, 8 active+clean, 36 unknown; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:17.406 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:17 vm03 bash[17055]: cluster 2026-03-06T21:20:16.085563+0000 mgr.vm03.uwuzgl (mgr.14199) 162 : cluster [DBG] pgmap v112: 65 pgs: 5 creating+activating, 16 creating+peering, 8 active+clean, 36 unknown; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:17.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:17 vm08 bash[20828]: cluster 2026-03-06T21:20:16.068519+0000 mon.vm03 (mon.0) 657 : cluster [DBG] osdmap e29: 8 total, 8 up, 8 in 2026-03-06T22:20:17.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:17 vm08 bash[20828]: cluster 2026-03-06T21:20:16.068519+0000 mon.vm03 (mon.0) 657 : cluster [DBG] osdmap e29: 8 total, 8 up, 8 in 2026-03-06T22:20:17.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:17 vm08 bash[20828]: cluster 2026-03-06T21:20:16.085563+0000 mgr.vm03.uwuzgl (mgr.14199) 162 : cluster [DBG] pgmap v112: 65 pgs: 5 creating+activating, 16 creating+peering, 8 active+clean, 36 unknown; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:17.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:17 vm08 bash[20828]: cluster 2026-03-06T21:20:16.085563+0000 mgr.vm03.uwuzgl (mgr.14199) 162 : cluster [DBG] pgmap v112: 65 pgs: 5 creating+activating, 16 creating+peering, 8 active+clean, 36 unknown; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:19.365 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: cluster 2026-03-06T21:20:18.086005+0000 mgr.vm03.uwuzgl (mgr.14199) 163 : cluster [DBG] pgmap v113: 65 pgs: 5 creating+activating, 16 creating+peering, 32 active+clean, 12 unknown; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: cluster 2026-03-06T21:20:18.086005+0000 mgr.vm03.uwuzgl (mgr.14199) 163 : cluster [DBG] pgmap v113: 65 pgs: 5 creating+activating, 16 creating+peering, 32 active+clean, 12 unknown; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.381969+0000 mon.vm03 (mon.0) 658 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.381969+0000 mon.vm03 (mon.0) 658 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.387879+0000 mon.vm03 (mon.0) 659 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.387879+0000 mon.vm03 (mon.0) 659 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.894953+0000 mon.vm03 (mon.0) 660 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.894953+0000 mon.vm03 (mon.0) 660 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.900886+0000 mon.vm03 (mon.0) 661 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.900886+0000 mon.vm03 (mon.0) 661 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.902069+0000 mon.vm03 (mon.0) 662 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.902069+0000 mon.vm03 (mon.0) 662 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.902690+0000 mon.vm03 (mon.0) 663 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.902690+0000 mon.vm03 (mon.0) 663 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.906773+0000 mon.vm03 (mon.0) 664 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.906773+0000 mon.vm03 (mon.0) 664 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.908346+0000 mon.vm03 (mon.0) 665 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.908346+0000 mon.vm03 (mon.0) 665 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.911452+0000 mon.vm03 (mon.0) 666 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mds.fs1.vm03.slersa", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.911452+0000 mon.vm03 (mon.0) 666 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mds.fs1.vm03.slersa", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.913748+0000 mon.vm03 (mon.0) 667 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "mds.fs1.vm03.slersa", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.913748+0000 mon.vm03 (mon.0) 667 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "mds.fs1.vm03.slersa", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.916055+0000 mon.vm03 (mon.0) 668 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:19.438 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 bash[17055]: audit 2026-03-06T21:20:18.916055+0000 mon.vm03 (mon.0) 668 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:19.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:20:19.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:19 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: cluster 2026-03-06T21:20:18.086005+0000 mgr.vm03.uwuzgl (mgr.14199) 163 : cluster [DBG] pgmap v113: 65 pgs: 5 creating+activating, 16 creating+peering, 32 active+clean, 12 unknown; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: cluster 2026-03-06T21:20:18.086005+0000 mgr.vm03.uwuzgl (mgr.14199) 163 : cluster [DBG] pgmap v113: 65 pgs: 5 creating+activating, 16 creating+peering, 32 active+clean, 12 unknown; 449 KiB data, 613 MiB used, 159 GiB / 160 GiB avail 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.381969+0000 mon.vm03 (mon.0) 658 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.381969+0000 mon.vm03 (mon.0) 658 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.387879+0000 mon.vm03 (mon.0) 659 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.387879+0000 mon.vm03 (mon.0) 659 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.894953+0000 mon.vm03 (mon.0) 660 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.894953+0000 mon.vm03 (mon.0) 660 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.900886+0000 mon.vm03 (mon.0) 661 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.900886+0000 mon.vm03 (mon.0) 661 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.902069+0000 mon.vm03 (mon.0) 662 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.902069+0000 mon.vm03 (mon.0) 662 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.902690+0000 mon.vm03 (mon.0) 663 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.902690+0000 mon.vm03 (mon.0) 663 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.906773+0000 mon.vm03 (mon.0) 664 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.906773+0000 mon.vm03 (mon.0) 664 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.908346+0000 mon.vm03 (mon.0) 665 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.908346+0000 mon.vm03 (mon.0) 665 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.911452+0000 mon.vm03 (mon.0) 666 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mds.fs1.vm03.slersa", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.911452+0000 mon.vm03 (mon.0) 666 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mds.fs1.vm03.slersa", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.913748+0000 mon.vm03 (mon.0) 667 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "mds.fs1.vm03.slersa", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.913748+0000 mon.vm03 (mon.0) 667 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "mds.fs1.vm03.slersa", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.916055+0000 mon.vm03 (mon.0) 668 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:19.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:19 vm08 bash[20828]: audit 2026-03-06T21:20:18.916055+0000 mon.vm03 (mon.0) 668 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:20.590 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:20:20.590 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 bash[20828]: cephadm 2026-03-06T21:20:18.916612+0000 mgr.vm03.uwuzgl (mgr.14199) 164 : cephadm [INF] Deploying daemon mds.fs1.vm03.slersa on vm03 2026-03-06T22:20:20.590 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 bash[20828]: cephadm 2026-03-06T21:20:18.916612+0000 mgr.vm03.uwuzgl (mgr.14199) 164 : cephadm [INF] Deploying daemon mds.fs1.vm03.slersa on vm03 2026-03-06T22:20:20.590 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 bash[20828]: audit 2026-03-06T21:20:19.767096+0000 mon.vm03 (mon.0) 669 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:20.590 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 bash[20828]: audit 2026-03-06T21:20:19.767096+0000 mon.vm03 (mon.0) 669 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:20.590 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 bash[20828]: audit 2026-03-06T21:20:19.774498+0000 mon.vm03 (mon.0) 670 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:20.590 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 bash[20828]: audit 2026-03-06T21:20:19.774498+0000 mon.vm03 (mon.0) 670 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:20.590 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 bash[20828]: audit 2026-03-06T21:20:19.779391+0000 mon.vm03 (mon.0) 671 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:20.590 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 bash[20828]: audit 2026-03-06T21:20:19.779391+0000 mon.vm03 (mon.0) 671 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:20.591 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 bash[20828]: audit 2026-03-06T21:20:19.780425+0000 mon.vm03 (mon.0) 672 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mds.fs1.vm08.gxmctn", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-06T22:20:20.591 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 bash[20828]: audit 2026-03-06T21:20:19.780425+0000 mon.vm03 (mon.0) 672 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mds.fs1.vm08.gxmctn", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-06T22:20:20.591 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 bash[20828]: audit 2026-03-06T21:20:19.785028+0000 mon.vm03 (mon.0) 673 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "mds.fs1.vm08.gxmctn", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-06T22:20:20.591 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 bash[20828]: audit 2026-03-06T21:20:19.785028+0000 mon.vm03 (mon.0) 673 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "mds.fs1.vm08.gxmctn", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-06T22:20:20.591 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 bash[20828]: audit 2026-03-06T21:20:19.787191+0000 mon.vm03 (mon.0) 674 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:20.591 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 bash[20828]: audit 2026-03-06T21:20:19.787191+0000 mon.vm03 (mon.0) 674 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:20.591 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 bash[20828]: audit 2026-03-06T21:20:20.082504+0000 mon.vm03 (mon.0) 675 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool create", "pool": ".nfs", "yes_i_really_mean_it": true}]: dispatch 2026-03-06T22:20:20.591 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 bash[20828]: audit 2026-03-06T21:20:20.082504+0000 mon.vm03 (mon.0) 675 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool create", "pool": ".nfs", "yes_i_really_mean_it": true}]: dispatch 2026-03-06T22:20:20.591 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:20 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:20:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:20 vm03 bash[17055]: cephadm 2026-03-06T21:20:18.916612+0000 mgr.vm03.uwuzgl (mgr.14199) 164 : cephadm [INF] Deploying daemon mds.fs1.vm03.slersa on vm03 2026-03-06T22:20:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:20 vm03 bash[17055]: cephadm 2026-03-06T21:20:18.916612+0000 mgr.vm03.uwuzgl (mgr.14199) 164 : cephadm [INF] Deploying daemon mds.fs1.vm03.slersa on vm03 2026-03-06T22:20:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:20 vm03 bash[17055]: audit 2026-03-06T21:20:19.767096+0000 mon.vm03 (mon.0) 669 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:20 vm03 bash[17055]: audit 2026-03-06T21:20:19.767096+0000 mon.vm03 (mon.0) 669 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:20 vm03 bash[17055]: audit 2026-03-06T21:20:19.774498+0000 mon.vm03 (mon.0) 670 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:20 vm03 bash[17055]: audit 2026-03-06T21:20:19.774498+0000 mon.vm03 (mon.0) 670 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:20 vm03 bash[17055]: audit 2026-03-06T21:20:19.779391+0000 mon.vm03 (mon.0) 671 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:20 vm03 bash[17055]: audit 2026-03-06T21:20:19.779391+0000 mon.vm03 (mon.0) 671 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:20 vm03 bash[17055]: audit 2026-03-06T21:20:19.780425+0000 mon.vm03 (mon.0) 672 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mds.fs1.vm08.gxmctn", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-06T22:20:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:20 vm03 bash[17055]: audit 2026-03-06T21:20:19.780425+0000 mon.vm03 (mon.0) 672 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "mds.fs1.vm08.gxmctn", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-06T22:20:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:20 vm03 bash[17055]: audit 2026-03-06T21:20:19.785028+0000 mon.vm03 (mon.0) 673 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "mds.fs1.vm08.gxmctn", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-06T22:20:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:20 vm03 bash[17055]: audit 2026-03-06T21:20:19.785028+0000 mon.vm03 (mon.0) 673 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "mds.fs1.vm08.gxmctn", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-06T22:20:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:20 vm03 bash[17055]: audit 2026-03-06T21:20:19.787191+0000 mon.vm03 (mon.0) 674 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:20 vm03 bash[17055]: audit 2026-03-06T21:20:19.787191+0000 mon.vm03 (mon.0) 674 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:20 vm03 bash[17055]: audit 2026-03-06T21:20:20.082504+0000 mon.vm03 (mon.0) 675 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool create", "pool": ".nfs", "yes_i_really_mean_it": true}]: dispatch 2026-03-06T22:20:20.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:20 vm03 bash[17055]: audit 2026-03-06T21:20:20.082504+0000 mon.vm03 (mon.0) 675 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool create", "pool": ".nfs", "yes_i_really_mean_it": true}]: dispatch 2026-03-06T22:20:21.397 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cephadm 2026-03-06T21:20:19.788666+0000 mgr.vm03.uwuzgl (mgr.14199) 165 : cephadm [INF] Deploying daemon mds.fs1.vm08.gxmctn on vm08 2026-03-06T22:20:21.397 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cephadm 2026-03-06T21:20:19.788666+0000 mgr.vm03.uwuzgl (mgr.14199) 165 : cephadm [INF] Deploying daemon mds.fs1.vm08.gxmctn on vm08 2026-03-06T22:20:21.397 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.081309+0000 mgr.vm03.uwuzgl (mgr.14199) 166 : audit [DBG] from='client.14478 -' entity='client.admin' cmd=[{"prefix": "nfs cluster create", "cluster_id": "happy", "ingress": true, "virtual_ip": "12.12.1.103", "ingress_mode": "haproxy-protocol", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:21.397 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.081309+0000 mgr.vm03.uwuzgl (mgr.14199) 166 : audit [DBG] from='client.14478 -' entity='client.admin' cmd=[{"prefix": "nfs cluster create", "cluster_id": "happy", "ingress": true, "virtual_ip": "12.12.1.103", "ingress_mode": "haproxy-protocol", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:21.397 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.086348+0000 mgr.vm03.uwuzgl (mgr.14199) 167 : cluster [DBG] pgmap v114: 65 pgs: 5 creating+activating, 60 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T22:20:21.397 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.086348+0000 mgr.vm03.uwuzgl (mgr.14199) 167 : cluster [DBG] pgmap v114: 65 pgs: 5 creating+activating, 60 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T22:20:21.397 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.612879+0000 mon.vm03 (mon.0) 676 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.397 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.612879+0000 mon.vm03 (mon.0) 676 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.397 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.618090+0000 mon.vm03 (mon.0) 677 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.397 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.618090+0000 mon.vm03 (mon.0) 677 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cephadm 2026-03-06T21:20:19.788666+0000 mgr.vm03.uwuzgl (mgr.14199) 165 : cephadm [INF] Deploying daemon mds.fs1.vm08.gxmctn on vm08 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cephadm 2026-03-06T21:20:19.788666+0000 mgr.vm03.uwuzgl (mgr.14199) 165 : cephadm [INF] Deploying daemon mds.fs1.vm08.gxmctn on vm08 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.081309+0000 mgr.vm03.uwuzgl (mgr.14199) 166 : audit [DBG] from='client.14478 -' entity='client.admin' cmd=[{"prefix": "nfs cluster create", "cluster_id": "happy", "ingress": true, "virtual_ip": "12.12.1.103", "ingress_mode": "haproxy-protocol", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.081309+0000 mgr.vm03.uwuzgl (mgr.14199) 166 : audit [DBG] from='client.14478 -' entity='client.admin' cmd=[{"prefix": "nfs cluster create", "cluster_id": "happy", "ingress": true, "virtual_ip": "12.12.1.103", "ingress_mode": "haproxy-protocol", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.086348+0000 mgr.vm03.uwuzgl (mgr.14199) 167 : cluster [DBG] pgmap v114: 65 pgs: 5 creating+activating, 60 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.086348+0000 mgr.vm03.uwuzgl (mgr.14199) 167 : cluster [DBG] pgmap v114: 65 pgs: 5 creating+activating, 60 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.612879+0000 mon.vm03 (mon.0) 676 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.612879+0000 mon.vm03 (mon.0) 676 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.618090+0000 mon.vm03 (mon.0) 677 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.618090+0000 mon.vm03 (mon.0) 677 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.626073+0000 mon.vm03 (mon.0) 678 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.626073+0000 mon.vm03 (mon.0) 678 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.629926+0000 mon.vm03 (mon.0) 679 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.629926+0000 mon.vm03 (mon.0) 679 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.634598+0000 mon.vm03 (mon.0) 680 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.634598+0000 mon.vm03 (mon.0) 680 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.644658+0000 mon.vm03 (mon.0) 681 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.644658+0000 mon.vm03 (mon.0) 681 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.792970+0000 mon.vm03 (mon.0) 682 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool create", "pool": ".nfs", "yes_i_really_mean_it": true}]': finished 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.792970+0000 mon.vm03 (mon.0) 682 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool create", "pool": ".nfs", "yes_i_really_mean_it": true}]': finished 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.795401+0000 mon.vm03 (mon.0) 683 : cluster [DBG] mds.? [v2:192.168.123.103:6834/3534785725,v1:192.168.123.103:6835/3534785725] up:boot 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.795401+0000 mon.vm03 (mon.0) 683 : cluster [DBG] mds.? [v2:192.168.123.103:6834/3534785725,v1:192.168.123.103:6835/3534785725] up:boot 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.795526+0000 mon.vm03 (mon.0) 684 : cluster [DBG] mds.? [v2:192.168.123.108:6832/402059897,v1:192.168.123.108:6833/402059897] up:boot 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.795526+0000 mon.vm03 (mon.0) 684 : cluster [DBG] mds.? [v2:192.168.123.108:6832/402059897,v1:192.168.123.108:6833/402059897] up:boot 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.795566+0000 mon.vm03 (mon.0) 685 : cluster [INF] daemon mds.fs1.vm08.gxmctn assigned to filesystem fs1 as rank 0 (now has 1 ranks) 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.795566+0000 mon.vm03 (mon.0) 685 : cluster [INF] daemon mds.fs1.vm08.gxmctn assigned to filesystem fs1 as rank 0 (now has 1 ranks) 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.795597+0000 mon.vm03 (mon.0) 686 : cluster [INF] Health check cleared: MDS_ALL_DOWN (was: 1 filesystem is offline) 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.795597+0000 mon.vm03 (mon.0) 686 : cluster [INF] Health check cleared: MDS_ALL_DOWN (was: 1 filesystem is offline) 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.795608+0000 mon.vm03 (mon.0) 687 : cluster [INF] Health check cleared: MDS_UP_LESS_THAN_MAX (was: 1 filesystem is online with fewer MDS than max_mds) 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.795608+0000 mon.vm03 (mon.0) 687 : cluster [INF] Health check cleared: MDS_UP_LESS_THAN_MAX (was: 1 filesystem is online with fewer MDS than max_mds) 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.795618+0000 mon.vm03 (mon.0) 688 : cluster [INF] Cluster is now healthy 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.795618+0000 mon.vm03 (mon.0) 688 : cluster [INF] Cluster is now healthy 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.796525+0000 mon.vm03 (mon.0) 689 : cluster [DBG] fsmap fs1:0 2 up:standby 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.796525+0000 mon.vm03 (mon.0) 689 : cluster [DBG] fsmap fs1:0 2 up:standby 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.796550+0000 mon.vm03 (mon.0) 690 : cluster [DBG] osdmap e30: 8 total, 8 up, 8 in 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.796550+0000 mon.vm03 (mon.0) 690 : cluster [DBG] osdmap e30: 8 total, 8 up, 8 in 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.801100+0000 mon.vm03 (mon.0) 691 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata", "who": "fs1.vm03.slersa"}]: dispatch 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.801100+0000 mon.vm03 (mon.0) 691 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata", "who": "fs1.vm03.slersa"}]: dispatch 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.801251+0000 mon.vm03 (mon.0) 692 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata", "who": "fs1.vm08.gxmctn"}]: dispatch 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.801251+0000 mon.vm03 (mon.0) 692 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata", "who": "fs1.vm08.gxmctn"}]: dispatch 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.822381+0000 mon.vm03 (mon.0) 693 : cluster [DBG] fsmap fs1:1 {0=fs1.vm08.gxmctn=up:creating} 1 up:standby 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.822381+0000 mon.vm03 (mon.0) 693 : cluster [DBG] fsmap fs1:1 {0=fs1.vm08.gxmctn=up:creating} 1 up:standby 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.822528+0000 mon.vm03 (mon.0) 694 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]: dispatch 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:20.822528+0000 mon.vm03 (mon.0) 694 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]: dispatch 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.876266+0000 mon.vm03 (mon.0) 695 : cluster [INF] daemon mds.fs1.vm08.gxmctn is now active in filesystem fs1 as rank 0 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.876266+0000 mon.vm03 (mon.0) 695 : cluster [INF] daemon mds.fs1.vm08.gxmctn is now active in filesystem fs1 as rank 0 2026-03-06T22:20:21.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.957491+0000 mon.vm03 (mon.0) 696 : cluster [DBG] mds.? [v2:192.168.123.108:6832/402059897,v1:192.168.123.108:6833/402059897] up:active 2026-03-06T22:20:21.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.957491+0000 mon.vm03 (mon.0) 696 : cluster [DBG] mds.? [v2:192.168.123.108:6832/402059897,v1:192.168.123.108:6833/402059897] up:active 2026-03-06T22:20:21.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.957700+0000 mon.vm03 (mon.0) 697 : cluster [DBG] fsmap fs1:1 {0=fs1.vm08.gxmctn=up:active} 1 up:standby 2026-03-06T22:20:21.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: cluster 2026-03-06T21:20:20.957700+0000 mon.vm03 (mon.0) 697 : cluster [DBG] fsmap fs1:1 {0=fs1.vm08.gxmctn=up:active} 1 up:standby 2026-03-06T22:20:21.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:21.058876+0000 mon.vm03 (mon.0) 698 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:20:21.692 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:21 vm03 bash[17055]: audit 2026-03-06T21:20:21.058876+0000 mon.vm03 (mon.0) 698 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:20:21.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.626073+0000 mon.vm03 (mon.0) 678 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.626073+0000 mon.vm03 (mon.0) 678 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.629926+0000 mon.vm03 (mon.0) 679 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.629926+0000 mon.vm03 (mon.0) 679 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.634598+0000 mon.vm03 (mon.0) 680 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.634598+0000 mon.vm03 (mon.0) 680 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.644658+0000 mon.vm03 (mon.0) 681 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.644658+0000 mon.vm03 (mon.0) 681 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.792970+0000 mon.vm03 (mon.0) 682 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool create", "pool": ".nfs", "yes_i_really_mean_it": true}]': finished 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.792970+0000 mon.vm03 (mon.0) 682 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool create", "pool": ".nfs", "yes_i_really_mean_it": true}]': finished 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.795401+0000 mon.vm03 (mon.0) 683 : cluster [DBG] mds.? [v2:192.168.123.103:6834/3534785725,v1:192.168.123.103:6835/3534785725] up:boot 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.795401+0000 mon.vm03 (mon.0) 683 : cluster [DBG] mds.? [v2:192.168.123.103:6834/3534785725,v1:192.168.123.103:6835/3534785725] up:boot 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.795526+0000 mon.vm03 (mon.0) 684 : cluster [DBG] mds.? [v2:192.168.123.108:6832/402059897,v1:192.168.123.108:6833/402059897] up:boot 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.795526+0000 mon.vm03 (mon.0) 684 : cluster [DBG] mds.? [v2:192.168.123.108:6832/402059897,v1:192.168.123.108:6833/402059897] up:boot 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.795566+0000 mon.vm03 (mon.0) 685 : cluster [INF] daemon mds.fs1.vm08.gxmctn assigned to filesystem fs1 as rank 0 (now has 1 ranks) 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.795566+0000 mon.vm03 (mon.0) 685 : cluster [INF] daemon mds.fs1.vm08.gxmctn assigned to filesystem fs1 as rank 0 (now has 1 ranks) 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.795597+0000 mon.vm03 (mon.0) 686 : cluster [INF] Health check cleared: MDS_ALL_DOWN (was: 1 filesystem is offline) 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.795597+0000 mon.vm03 (mon.0) 686 : cluster [INF] Health check cleared: MDS_ALL_DOWN (was: 1 filesystem is offline) 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.795608+0000 mon.vm03 (mon.0) 687 : cluster [INF] Health check cleared: MDS_UP_LESS_THAN_MAX (was: 1 filesystem is online with fewer MDS than max_mds) 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.795608+0000 mon.vm03 (mon.0) 687 : cluster [INF] Health check cleared: MDS_UP_LESS_THAN_MAX (was: 1 filesystem is online with fewer MDS than max_mds) 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.795618+0000 mon.vm03 (mon.0) 688 : cluster [INF] Cluster is now healthy 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.795618+0000 mon.vm03 (mon.0) 688 : cluster [INF] Cluster is now healthy 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.796525+0000 mon.vm03 (mon.0) 689 : cluster [DBG] fsmap fs1:0 2 up:standby 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.796525+0000 mon.vm03 (mon.0) 689 : cluster [DBG] fsmap fs1:0 2 up:standby 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.796550+0000 mon.vm03 (mon.0) 690 : cluster [DBG] osdmap e30: 8 total, 8 up, 8 in 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.796550+0000 mon.vm03 (mon.0) 690 : cluster [DBG] osdmap e30: 8 total, 8 up, 8 in 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.801100+0000 mon.vm03 (mon.0) 691 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata", "who": "fs1.vm03.slersa"}]: dispatch 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.801100+0000 mon.vm03 (mon.0) 691 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata", "who": "fs1.vm03.slersa"}]: dispatch 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.801251+0000 mon.vm03 (mon.0) 692 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata", "who": "fs1.vm08.gxmctn"}]: dispatch 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.801251+0000 mon.vm03 (mon.0) 692 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "mds metadata", "who": "fs1.vm08.gxmctn"}]: dispatch 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.822381+0000 mon.vm03 (mon.0) 693 : cluster [DBG] fsmap fs1:1 {0=fs1.vm08.gxmctn=up:creating} 1 up:standby 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.822381+0000 mon.vm03 (mon.0) 693 : cluster [DBG] fsmap fs1:1 {0=fs1.vm08.gxmctn=up:creating} 1 up:standby 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.822528+0000 mon.vm03 (mon.0) 694 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]: dispatch 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:20.822528+0000 mon.vm03 (mon.0) 694 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]: dispatch 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.876266+0000 mon.vm03 (mon.0) 695 : cluster [INF] daemon mds.fs1.vm08.gxmctn is now active in filesystem fs1 as rank 0 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.876266+0000 mon.vm03 (mon.0) 695 : cluster [INF] daemon mds.fs1.vm08.gxmctn is now active in filesystem fs1 as rank 0 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.957491+0000 mon.vm03 (mon.0) 696 : cluster [DBG] mds.? [v2:192.168.123.108:6832/402059897,v1:192.168.123.108:6833/402059897] up:active 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.957491+0000 mon.vm03 (mon.0) 696 : cluster [DBG] mds.? [v2:192.168.123.108:6832/402059897,v1:192.168.123.108:6833/402059897] up:active 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.957700+0000 mon.vm03 (mon.0) 697 : cluster [DBG] fsmap fs1:1 {0=fs1.vm08.gxmctn=up:active} 1 up:standby 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: cluster 2026-03-06T21:20:20.957700+0000 mon.vm03 (mon.0) 697 : cluster [DBG] fsmap fs1:1 {0=fs1.vm08.gxmctn=up:active} 1 up:standby 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:21.058876+0000 mon.vm03 (mon.0) 698 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:20:21.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:21 vm08 bash[20828]: audit 2026-03-06T21:20:21.058876+0000 mon.vm03 (mon.0) 698 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:20:21.928 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- bash -c 'ceph nfs export create cephfs --fsname fs1 --cluster-id happy --pseudo-path /d1' 2026-03-06T22:20:23.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:22 vm03 bash[17055]: audit 2026-03-06T21:20:21.796511+0000 mon.vm03 (mon.0) 699 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]': finished 2026-03-06T22:20:23.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:22 vm03 bash[17055]: audit 2026-03-06T21:20:21.796511+0000 mon.vm03 (mon.0) 699 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]': finished 2026-03-06T22:20:23.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:22 vm03 bash[17055]: cluster 2026-03-06T21:20:21.801419+0000 mon.vm03 (mon.0) 700 : cluster [DBG] osdmap e31: 8 total, 8 up, 8 in 2026-03-06T22:20:23.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:22 vm03 bash[17055]: cluster 2026-03-06T21:20:21.801419+0000 mon.vm03 (mon.0) 700 : cluster [DBG] osdmap e31: 8 total, 8 up, 8 in 2026-03-06T22:20:23.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:22 vm03 bash[17055]: cephadm 2026-03-06T21:20:21.842596+0000 mgr.vm03.uwuzgl (mgr.14199) 168 : cephadm [INF] Saving service nfs.happy spec with placement count:1 2026-03-06T22:20:23.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:22 vm03 bash[17055]: cephadm 2026-03-06T21:20:21.842596+0000 mgr.vm03.uwuzgl (mgr.14199) 168 : cephadm [INF] Saving service nfs.happy spec with placement count:1 2026-03-06T22:20:23.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:22 vm03 bash[17055]: audit 2026-03-06T21:20:21.846171+0000 mon.vm03 (mon.0) 701 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:23.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:22 vm03 bash[17055]: audit 2026-03-06T21:20:21.846171+0000 mon.vm03 (mon.0) 701 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:23.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:22 vm03 bash[17055]: cephadm 2026-03-06T21:20:21.847170+0000 mgr.vm03.uwuzgl (mgr.14199) 169 : cephadm [INF] Saving service ingress.nfs.happy spec with placement count:2 2026-03-06T22:20:23.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:22 vm03 bash[17055]: cephadm 2026-03-06T21:20:21.847170+0000 mgr.vm03.uwuzgl (mgr.14199) 169 : cephadm [INF] Saving service ingress.nfs.happy spec with placement count:2 2026-03-06T22:20:23.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:22 vm03 bash[17055]: audit 2026-03-06T21:20:21.850594+0000 mon.vm03 (mon.0) 702 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:23.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:22 vm03 bash[17055]: audit 2026-03-06T21:20:21.850594+0000 mon.vm03 (mon.0) 702 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:23.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:22 vm03 bash[17055]: cluster 2026-03-06T21:20:22.086617+0000 mgr.vm03.uwuzgl (mgr.14199) 170 : cluster [DBG] pgmap v117: 97 pgs: 32 unknown, 65 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T22:20:23.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:22 vm03 bash[17055]: cluster 2026-03-06T21:20:22.086617+0000 mgr.vm03.uwuzgl (mgr.14199) 170 : cluster [DBG] pgmap v117: 97 pgs: 32 unknown, 65 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T22:20:23.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:22 vm03 bash[17055]: audit 2026-03-06T21:20:22.551711+0000 mon.vm03 (mon.0) 703 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:23.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:22 vm03 bash[17055]: audit 2026-03-06T21:20:22.551711+0000 mon.vm03 (mon.0) 703 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:23.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:22 vm08 bash[20828]: audit 2026-03-06T21:20:21.796511+0000 mon.vm03 (mon.0) 699 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]': finished 2026-03-06T22:20:23.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:22 vm08 bash[20828]: audit 2026-03-06T21:20:21.796511+0000 mon.vm03 (mon.0) 699 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]': finished 2026-03-06T22:20:23.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:22 vm08 bash[20828]: cluster 2026-03-06T21:20:21.801419+0000 mon.vm03 (mon.0) 700 : cluster [DBG] osdmap e31: 8 total, 8 up, 8 in 2026-03-06T22:20:23.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:22 vm08 bash[20828]: cluster 2026-03-06T21:20:21.801419+0000 mon.vm03 (mon.0) 700 : cluster [DBG] osdmap e31: 8 total, 8 up, 8 in 2026-03-06T22:20:23.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:22 vm08 bash[20828]: cephadm 2026-03-06T21:20:21.842596+0000 mgr.vm03.uwuzgl (mgr.14199) 168 : cephadm [INF] Saving service nfs.happy spec with placement count:1 2026-03-06T22:20:23.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:22 vm08 bash[20828]: cephadm 2026-03-06T21:20:21.842596+0000 mgr.vm03.uwuzgl (mgr.14199) 168 : cephadm [INF] Saving service nfs.happy spec with placement count:1 2026-03-06T22:20:23.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:22 vm08 bash[20828]: audit 2026-03-06T21:20:21.846171+0000 mon.vm03 (mon.0) 701 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:23.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:22 vm08 bash[20828]: audit 2026-03-06T21:20:21.846171+0000 mon.vm03 (mon.0) 701 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:23.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:22 vm08 bash[20828]: cephadm 2026-03-06T21:20:21.847170+0000 mgr.vm03.uwuzgl (mgr.14199) 169 : cephadm [INF] Saving service ingress.nfs.happy spec with placement count:2 2026-03-06T22:20:23.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:22 vm08 bash[20828]: cephadm 2026-03-06T21:20:21.847170+0000 mgr.vm03.uwuzgl (mgr.14199) 169 : cephadm [INF] Saving service ingress.nfs.happy spec with placement count:2 2026-03-06T22:20:23.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:22 vm08 bash[20828]: audit 2026-03-06T21:20:21.850594+0000 mon.vm03 (mon.0) 702 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:23.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:22 vm08 bash[20828]: audit 2026-03-06T21:20:21.850594+0000 mon.vm03 (mon.0) 702 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:23.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:22 vm08 bash[20828]: cluster 2026-03-06T21:20:22.086617+0000 mgr.vm03.uwuzgl (mgr.14199) 170 : cluster [DBG] pgmap v117: 97 pgs: 32 unknown, 65 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T22:20:23.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:22 vm08 bash[20828]: cluster 2026-03-06T21:20:22.086617+0000 mgr.vm03.uwuzgl (mgr.14199) 170 : cluster [DBG] pgmap v117: 97 pgs: 32 unknown, 65 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-06T22:20:23.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:22 vm08 bash[20828]: audit 2026-03-06T21:20:22.551711+0000 mon.vm03 (mon.0) 703 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:23.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:22 vm08 bash[20828]: audit 2026-03-06T21:20:22.551711+0000 mon.vm03 (mon.0) 703 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:24.128 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:23 vm08 bash[20828]: cluster 2026-03-06T21:20:22.803734+0000 mon.vm03 (mon.0) 704 : cluster [DBG] osdmap e32: 8 total, 8 up, 8 in 2026-03-06T22:20:24.128 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:23 vm08 bash[20828]: cluster 2026-03-06T21:20:22.803734+0000 mon.vm03 (mon.0) 704 : cluster [DBG] osdmap e32: 8 total, 8 up, 8 in 2026-03-06T22:20:24.128 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:23 vm08 bash[20828]: cluster 2026-03-06T21:20:22.871211+0000 mon.vm03 (mon.0) 705 : cluster [DBG] mgrmap e20: vm03.uwuzgl(active, since 3m), standbys: vm08.tdhano 2026-03-06T22:20:24.128 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:23 vm08 bash[20828]: cluster 2026-03-06T21:20:22.871211+0000 mon.vm03 (mon.0) 705 : cluster [DBG] mgrmap e20: vm03.uwuzgl(active, since 3m), standbys: vm08.tdhano 2026-03-06T22:20:24.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:23 vm03 bash[17055]: cluster 2026-03-06T21:20:22.803734+0000 mon.vm03 (mon.0) 704 : cluster [DBG] osdmap e32: 8 total, 8 up, 8 in 2026-03-06T22:20:24.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:23 vm03 bash[17055]: cluster 2026-03-06T21:20:22.803734+0000 mon.vm03 (mon.0) 704 : cluster [DBG] osdmap e32: 8 total, 8 up, 8 in 2026-03-06T22:20:24.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:23 vm03 bash[17055]: cluster 2026-03-06T21:20:22.871211+0000 mon.vm03 (mon.0) 705 : cluster [DBG] mgrmap e20: vm03.uwuzgl(active, since 3m), standbys: vm08.tdhano 2026-03-06T22:20:24.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:23 vm03 bash[17055]: cluster 2026-03-06T21:20:22.871211+0000 mon.vm03 (mon.0) 705 : cluster [DBG] mgrmap e20: vm03.uwuzgl(active, since 3m), standbys: vm08.tdhano 2026-03-06T22:20:25.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:24 vm03 bash[17055]: cluster 2026-03-06T21:20:24.087041+0000 mgr.vm03.uwuzgl (mgr.14199) 171 : cluster [DBG] pgmap v119: 97 pgs: 6 unknown, 91 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s wr, 3 op/s 2026-03-06T22:20:25.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:24 vm03 bash[17055]: cluster 2026-03-06T21:20:24.087041+0000 mgr.vm03.uwuzgl (mgr.14199) 171 : cluster [DBG] pgmap v119: 97 pgs: 6 unknown, 91 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s wr, 3 op/s 2026-03-06T22:20:25.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:24 vm08 bash[20828]: cluster 2026-03-06T21:20:24.087041+0000 mgr.vm03.uwuzgl (mgr.14199) 171 : cluster [DBG] pgmap v119: 97 pgs: 6 unknown, 91 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s wr, 3 op/s 2026-03-06T22:20:25.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:24 vm08 bash[20828]: cluster 2026-03-06T21:20:24.087041+0000 mgr.vm03.uwuzgl (mgr.14199) 171 : cluster [DBG] pgmap v119: 97 pgs: 6 unknown, 91 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s wr, 3 op/s 2026-03-06T22:20:26.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:25 vm03 bash[17055]: cluster 2026-03-06T21:20:24.821171+0000 mon.vm03 (mon.0) 706 : cluster [DBG] mds.? [v2:192.168.123.103:6834/3534785725,v1:192.168.123.103:6835/3534785725] up:standby 2026-03-06T22:20:26.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:25 vm03 bash[17055]: cluster 2026-03-06T21:20:24.821171+0000 mon.vm03 (mon.0) 706 : cluster [DBG] mds.? [v2:192.168.123.103:6834/3534785725,v1:192.168.123.103:6835/3534785725] up:standby 2026-03-06T22:20:26.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:25 vm03 bash[17055]: cluster 2026-03-06T21:20:24.821406+0000 mon.vm03 (mon.0) 707 : cluster [DBG] fsmap fs1:1 {0=fs1.vm08.gxmctn=up:active} 1 up:standby 2026-03-06T22:20:26.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:25 vm03 bash[17055]: cluster 2026-03-06T21:20:24.821406+0000 mon.vm03 (mon.0) 707 : cluster [DBG] fsmap fs1:1 {0=fs1.vm08.gxmctn=up:active} 1 up:standby 2026-03-06T22:20:26.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:25 vm08 bash[20828]: cluster 2026-03-06T21:20:24.821171+0000 mon.vm03 (mon.0) 706 : cluster [DBG] mds.? [v2:192.168.123.103:6834/3534785725,v1:192.168.123.103:6835/3534785725] up:standby 2026-03-06T22:20:26.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:25 vm08 bash[20828]: cluster 2026-03-06T21:20:24.821171+0000 mon.vm03 (mon.0) 706 : cluster [DBG] mds.? [v2:192.168.123.103:6834/3534785725,v1:192.168.123.103:6835/3534785725] up:standby 2026-03-06T22:20:26.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:25 vm08 bash[20828]: cluster 2026-03-06T21:20:24.821406+0000 mon.vm03 (mon.0) 707 : cluster [DBG] fsmap fs1:1 {0=fs1.vm08.gxmctn=up:active} 1 up:standby 2026-03-06T22:20:26.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:25 vm08 bash[20828]: cluster 2026-03-06T21:20:24.821406+0000 mon.vm03 (mon.0) 707 : cluster [DBG] fsmap fs1:1 {0=fs1.vm08.gxmctn=up:active} 1 up:standby 2026-03-06T22:20:26.488 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:20:26.910 INFO:teuthology.orchestra.run.vm03.stdout:{ 2026-03-06T22:20:26.910 INFO:teuthology.orchestra.run.vm03.stdout: "bind": "/d1", 2026-03-06T22:20:26.910 INFO:teuthology.orchestra.run.vm03.stdout: "cluster": "happy", 2026-03-06T22:20:26.910 INFO:teuthology.orchestra.run.vm03.stdout: "fs": "fs1", 2026-03-06T22:20:26.910 INFO:teuthology.orchestra.run.vm03.stdout: "mode": "RW", 2026-03-06T22:20:26.910 INFO:teuthology.orchestra.run.vm03.stdout: "path": "/" 2026-03-06T22:20:26.910 INFO:teuthology.orchestra.run.vm03.stdout:} 2026-03-06T22:20:26.981 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-06T22:20:26.995 INFO:tasks.cephadm:Waiting for ceph service nfs.happy to start (timeout 300)... 2026-03-06T22:20:26.995 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph orch ls -f json 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:25.842726+0000 mon.vm03 (mon.0) 708 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:25.842726+0000 mon.vm03 (mon.0) 708 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:25.849043+0000 mon.vm03 (mon.0) 709 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:25.849043+0000 mon.vm03 (mon.0) 709 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: cluster 2026-03-06T21:20:26.087376+0000 mgr.vm03.uwuzgl (mgr.14199) 172 : cluster [DBG] pgmap v120: 97 pgs: 97 active+clean; 451 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s wr, 7 op/s 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: cluster 2026-03-06T21:20:26.087376+0000 mgr.vm03.uwuzgl (mgr.14199) 172 : cluster [DBG] pgmap v120: 97 pgs: 97 active+clean; 451 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s wr, 7 op/s 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.096277+0000 mon.vm03 (mon.0) 710 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.096277+0000 mon.vm03 (mon.0) 710 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.101343+0000 mon.vm03 (mon.0) 711 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.101343+0000 mon.vm03 (mon.0) 711 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.102216+0000 mon.vm03 (mon.0) 712 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.102216+0000 mon.vm03 (mon.0) 712 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.102674+0000 mon.vm03 (mon.0) 713 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.102674+0000 mon.vm03 (mon.0) 713 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.106295+0000 mon.vm03 (mon.0) 714 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.106295+0000 mon.vm03 (mon.0) 714 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.107565+0000 mon.vm03 (mon.0) 715 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.107565+0000 mon.vm03 (mon.0) 715 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.113806+0000 mon.vm03 (mon.0) 716 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.113806+0000 mon.vm03 (mon.0) 716 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: cephadm 2026-03-06T21:20:26.114189+0000 mgr.vm03.uwuzgl (mgr.14199) 173 : cephadm [INF] Creating key for client.nfs.happy.0.0.vm08.wdijel 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: cephadm 2026-03-06T21:20:26.114189+0000 mgr.vm03.uwuzgl (mgr.14199) 173 : cephadm [INF] Creating key for client.nfs.happy.0.0.vm08.wdijel 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.114340+0000 mon.vm03 (mon.0) 717 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.0.0.vm08.wdijel", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy"]}]: dispatch 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.114340+0000 mon.vm03 (mon.0) 717 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.0.0.vm08.wdijel", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy"]}]: dispatch 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.116651+0000 mon.vm03 (mon.0) 718 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.0.0.vm08.wdijel", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy"]}]': finished 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.116651+0000 mon.vm03 (mon.0) 718 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.0.0.vm08.wdijel", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy"]}]': finished 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: cephadm 2026-03-06T21:20:26.118310+0000 mgr.vm03.uwuzgl (mgr.14199) 174 : cephadm [INF] Ensuring nfs.happy.0 is in the ganesha grace table 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: cephadm 2026-03-06T21:20:26.118310+0000 mgr.vm03.uwuzgl (mgr.14199) 174 : cephadm [INF] Ensuring nfs.happy.0 is in the ganesha grace table 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.118458+0000 mon.vm03 (mon.0) 719 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.happy", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.118458+0000 mon.vm03 (mon.0) 719 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.happy", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.119951+0000 mon.vm03 (mon.0) 720 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.happy", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.119951+0000 mon.vm03 (mon.0) 720 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.happy", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.121770+0000 mon.vm03 (mon.0) 721 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.121770+0000 mon.vm03 (mon.0) 721 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.159243+0000 mon.vm03 (mon.0) 722 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.happy"}]: dispatch 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.159243+0000 mon.vm03 (mon.0) 722 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.happy"}]: dispatch 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.161435+0000 mon.vm03 (mon.0) 723 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.happy"}]': finished 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.161435+0000 mon.vm03 (mon.0) 723 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.happy"}]': finished 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: cephadm 2026-03-06T21:20:26.202666+0000 mgr.vm03.uwuzgl (mgr.14199) 175 : cephadm [INF] Rados config object exists: conf-nfs.happy 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: cephadm 2026-03-06T21:20:26.202666+0000 mgr.vm03.uwuzgl (mgr.14199) 175 : cephadm [INF] Rados config object exists: conf-nfs.happy 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: cephadm 2026-03-06T21:20:26.202725+0000 mgr.vm03.uwuzgl (mgr.14199) 176 : cephadm [INF] Creating key for client.nfs.happy.0.0.vm08.wdijel-rgw 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: cephadm 2026-03-06T21:20:26.202725+0000 mgr.vm03.uwuzgl (mgr.14199) 176 : cephadm [INF] Creating key for client.nfs.happy.0.0.vm08.wdijel-rgw 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.203005+0000 mon.vm03 (mon.0) 724 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.0.0.vm08.wdijel-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.203005+0000 mon.vm03 (mon.0) 724 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.0.0.vm08.wdijel-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.205466+0000 mon.vm03 (mon.0) 725 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.0.0.vm08.wdijel-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.205466+0000 mon.vm03 (mon.0) 725 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.0.0.vm08.wdijel-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: cephadm 2026-03-06T21:20:26.207419+0000 mgr.vm03.uwuzgl (mgr.14199) 177 : cephadm [WRN] Bind address in nfs.happy.0.0.vm08.wdijel's ganesha conf is defaulting to empty 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: cephadm 2026-03-06T21:20:26.207419+0000 mgr.vm03.uwuzgl (mgr.14199) 177 : cephadm [WRN] Bind address in nfs.happy.0.0.vm08.wdijel's ganesha conf is defaulting to empty 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.210502+0000 mon.vm03 (mon.0) 726 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: audit 2026-03-06T21:20:26.210502+0000 mon.vm03 (mon.0) 726 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: cephadm 2026-03-06T21:20:26.211358+0000 mgr.vm03.uwuzgl (mgr.14199) 178 : cephadm [INF] Deploying daemon nfs.happy.0.0.vm08.wdijel on vm08 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 bash[20828]: cephadm 2026-03-06T21:20:26.211358+0000 mgr.vm03.uwuzgl (mgr.14199) 178 : cephadm [INF] Deploying daemon nfs.happy.0.0.vm08.wdijel on vm08 2026-03-06T22:20:27.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:26 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:25.842726+0000 mon.vm03 (mon.0) 708 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:25.842726+0000 mon.vm03 (mon.0) 708 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:25.849043+0000 mon.vm03 (mon.0) 709 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:25.849043+0000 mon.vm03 (mon.0) 709 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: cluster 2026-03-06T21:20:26.087376+0000 mgr.vm03.uwuzgl (mgr.14199) 172 : cluster [DBG] pgmap v120: 97 pgs: 97 active+clean; 451 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s wr, 7 op/s 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: cluster 2026-03-06T21:20:26.087376+0000 mgr.vm03.uwuzgl (mgr.14199) 172 : cluster [DBG] pgmap v120: 97 pgs: 97 active+clean; 451 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s wr, 7 op/s 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.096277+0000 mon.vm03 (mon.0) 710 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.096277+0000 mon.vm03 (mon.0) 710 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.101343+0000 mon.vm03 (mon.0) 711 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.101343+0000 mon.vm03 (mon.0) 711 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.102216+0000 mon.vm03 (mon.0) 712 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.102216+0000 mon.vm03 (mon.0) 712 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.102674+0000 mon.vm03 (mon.0) 713 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.102674+0000 mon.vm03 (mon.0) 713 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.106295+0000 mon.vm03 (mon.0) 714 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.106295+0000 mon.vm03 (mon.0) 714 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.107565+0000 mon.vm03 (mon.0) 715 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.107565+0000 mon.vm03 (mon.0) 715 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.113806+0000 mon.vm03 (mon.0) 716 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.113806+0000 mon.vm03 (mon.0) 716 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: cephadm 2026-03-06T21:20:26.114189+0000 mgr.vm03.uwuzgl (mgr.14199) 173 : cephadm [INF] Creating key for client.nfs.happy.0.0.vm08.wdijel 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: cephadm 2026-03-06T21:20:26.114189+0000 mgr.vm03.uwuzgl (mgr.14199) 173 : cephadm [INF] Creating key for client.nfs.happy.0.0.vm08.wdijel 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.114340+0000 mon.vm03 (mon.0) 717 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.0.0.vm08.wdijel", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy"]}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.114340+0000 mon.vm03 (mon.0) 717 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.0.0.vm08.wdijel", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy"]}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.116651+0000 mon.vm03 (mon.0) 718 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.0.0.vm08.wdijel", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy"]}]': finished 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.116651+0000 mon.vm03 (mon.0) 718 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.0.0.vm08.wdijel", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy"]}]': finished 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: cephadm 2026-03-06T21:20:26.118310+0000 mgr.vm03.uwuzgl (mgr.14199) 174 : cephadm [INF] Ensuring nfs.happy.0 is in the ganesha grace table 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: cephadm 2026-03-06T21:20:26.118310+0000 mgr.vm03.uwuzgl (mgr.14199) 174 : cephadm [INF] Ensuring nfs.happy.0 is in the ganesha grace table 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.118458+0000 mon.vm03 (mon.0) 719 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.happy", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.118458+0000 mon.vm03 (mon.0) 719 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.happy", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.119951+0000 mon.vm03 (mon.0) 720 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.happy", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.119951+0000 mon.vm03 (mon.0) 720 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.happy", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.121770+0000 mon.vm03 (mon.0) 721 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.121770+0000 mon.vm03 (mon.0) 721 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.159243+0000 mon.vm03 (mon.0) 722 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.happy"}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.159243+0000 mon.vm03 (mon.0) 722 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.happy"}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.161435+0000 mon.vm03 (mon.0) 723 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.happy"}]': finished 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.161435+0000 mon.vm03 (mon.0) 723 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.happy"}]': finished 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: cephadm 2026-03-06T21:20:26.202666+0000 mgr.vm03.uwuzgl (mgr.14199) 175 : cephadm [INF] Rados config object exists: conf-nfs.happy 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: cephadm 2026-03-06T21:20:26.202666+0000 mgr.vm03.uwuzgl (mgr.14199) 175 : cephadm [INF] Rados config object exists: conf-nfs.happy 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: cephadm 2026-03-06T21:20:26.202725+0000 mgr.vm03.uwuzgl (mgr.14199) 176 : cephadm [INF] Creating key for client.nfs.happy.0.0.vm08.wdijel-rgw 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: cephadm 2026-03-06T21:20:26.202725+0000 mgr.vm03.uwuzgl (mgr.14199) 176 : cephadm [INF] Creating key for client.nfs.happy.0.0.vm08.wdijel-rgw 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.203005+0000 mon.vm03 (mon.0) 724 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.0.0.vm08.wdijel-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.203005+0000 mon.vm03 (mon.0) 724 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.0.0.vm08.wdijel-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.205466+0000 mon.vm03 (mon.0) 725 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.0.0.vm08.wdijel-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.205466+0000 mon.vm03 (mon.0) 725 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.0.0.vm08.wdijel-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: cephadm 2026-03-06T21:20:26.207419+0000 mgr.vm03.uwuzgl (mgr.14199) 177 : cephadm [WRN] Bind address in nfs.happy.0.0.vm08.wdijel's ganesha conf is defaulting to empty 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: cephadm 2026-03-06T21:20:26.207419+0000 mgr.vm03.uwuzgl (mgr.14199) 177 : cephadm [WRN] Bind address in nfs.happy.0.0.vm08.wdijel's ganesha conf is defaulting to empty 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.210502+0000 mon.vm03 (mon.0) 726 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: audit 2026-03-06T21:20:26.210502+0000 mon.vm03 (mon.0) 726 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: cephadm 2026-03-06T21:20:26.211358+0000 mgr.vm03.uwuzgl (mgr.14199) 178 : cephadm [INF] Deploying daemon nfs.happy.0.0.vm08.wdijel on vm08 2026-03-06T22:20:27.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:26 vm03 bash[17055]: cephadm 2026-03-06T21:20:26.211358+0000 mgr.vm03.uwuzgl (mgr.14199) 178 : cephadm [INF] Deploying daemon nfs.happy.0.0.vm08.wdijel on vm08 2026-03-06T22:20:28.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:26.857891+0000 mgr.vm03.uwuzgl (mgr.14199) 179 : audit [DBG] from='client.14498 -' entity='client.admin' cmd=[{"prefix": "nfs export create cephfs", "fsname": "fs1", "cluster_id": "happy", "pseudo_path": "/d1", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:26.857891+0000 mgr.vm03.uwuzgl (mgr.14199) 179 : audit [DBG] from='client.14498 -' entity='client.admin' cmd=[{"prefix": "nfs export create cephfs", "fsname": "fs1", "cluster_id": "happy", "pseudo_path": "/d1", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:26.895164+0000 mon.vm03 (mon.0) 727 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.fs1.77683d3d", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy, allow rw tag cephfs data=fs1", "mds", "allow rw path=/"], "format": "json"}]: dispatch 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:26.895164+0000 mon.vm03 (mon.0) 727 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.fs1.77683d3d", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy, allow rw tag cephfs data=fs1", "mds", "allow rw path=/"], "format": "json"}]: dispatch 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:26.897292+0000 mon.vm03 (mon.0) 728 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.fs1.77683d3d", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy, allow rw tag cephfs data=fs1", "mds", "allow rw path=/"], "format": "json"}]': finished 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:26.897292+0000 mon.vm03 (mon.0) 728 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.fs1.77683d3d", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy, allow rw tag cephfs data=fs1", "mds", "allow rw path=/"], "format": "json"}]': finished 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:26.900242+0000 mon.vm03 (mon.0) 729 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.fs1.77683d3d", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy, allow rw tag cephfs data=fs1", "mds", "allow rw path=/"], "format": "json"}]: dispatch 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:26.900242+0000 mon.vm03 (mon.0) 729 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.fs1.77683d3d", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy, allow rw tag cephfs data=fs1", "mds", "allow rw path=/"], "format": "json"}]: dispatch 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:27.052661+0000 mon.vm03 (mon.0) 730 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:27.052661+0000 mon.vm03 (mon.0) 730 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:27.061204+0000 mon.vm03 (mon.0) 731 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:27.061204+0000 mon.vm03 (mon.0) 731 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:27.071707+0000 mon.vm03 (mon.0) 732 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:27.071707+0000 mon.vm03 (mon.0) 732 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:27.089671+0000 mon.vm03 (mon.0) 733 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:27.089671+0000 mon.vm03 (mon.0) 733 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:27.095537+0000 mon.vm03 (mon.0) 734 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:27.095537+0000 mon.vm03 (mon.0) 734 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: cephadm 2026-03-06T21:20:27.099981+0000 mgr.vm03.uwuzgl (mgr.14199) 180 : cephadm [INF] Deploying daemon haproxy.nfs.happy.vm08.kqaecw on vm08 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: cephadm 2026-03-06T21:20:27.099981+0000 mgr.vm03.uwuzgl (mgr.14199) 180 : cephadm [INF] Deploying daemon haproxy.nfs.happy.vm08.kqaecw on vm08 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:27.558833+0000 mon.vm03 (mon.0) 735 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.191 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:27 vm03 bash[17055]: audit 2026-03-06T21:20:27.558833+0000 mon.vm03 (mon.0) 735 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:26.857891+0000 mgr.vm03.uwuzgl (mgr.14199) 179 : audit [DBG] from='client.14498 -' entity='client.admin' cmd=[{"prefix": "nfs export create cephfs", "fsname": "fs1", "cluster_id": "happy", "pseudo_path": "/d1", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:26.857891+0000 mgr.vm03.uwuzgl (mgr.14199) 179 : audit [DBG] from='client.14498 -' entity='client.admin' cmd=[{"prefix": "nfs export create cephfs", "fsname": "fs1", "cluster_id": "happy", "pseudo_path": "/d1", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:26.895164+0000 mon.vm03 (mon.0) 727 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.fs1.77683d3d", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy, allow rw tag cephfs data=fs1", "mds", "allow rw path=/"], "format": "json"}]: dispatch 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:26.895164+0000 mon.vm03 (mon.0) 727 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.fs1.77683d3d", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy, allow rw tag cephfs data=fs1", "mds", "allow rw path=/"], "format": "json"}]: dispatch 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:26.897292+0000 mon.vm03 (mon.0) 728 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.fs1.77683d3d", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy, allow rw tag cephfs data=fs1", "mds", "allow rw path=/"], "format": "json"}]': finished 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:26.897292+0000 mon.vm03 (mon.0) 728 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.fs1.77683d3d", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy, allow rw tag cephfs data=fs1", "mds", "allow rw path=/"], "format": "json"}]': finished 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:26.900242+0000 mon.vm03 (mon.0) 729 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.fs1.77683d3d", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy, allow rw tag cephfs data=fs1", "mds", "allow rw path=/"], "format": "json"}]: dispatch 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:26.900242+0000 mon.vm03 (mon.0) 729 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.happy.fs1.77683d3d", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=happy, allow rw tag cephfs data=fs1", "mds", "allow rw path=/"], "format": "json"}]: dispatch 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:27.052661+0000 mon.vm03 (mon.0) 730 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:27.052661+0000 mon.vm03 (mon.0) 730 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:27.061204+0000 mon.vm03 (mon.0) 731 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:27.061204+0000 mon.vm03 (mon.0) 731 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:27.071707+0000 mon.vm03 (mon.0) 732 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:27.071707+0000 mon.vm03 (mon.0) 732 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:27.089671+0000 mon.vm03 (mon.0) 733 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:27.089671+0000 mon.vm03 (mon.0) 733 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:27.095537+0000 mon.vm03 (mon.0) 734 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:27.095537+0000 mon.vm03 (mon.0) 734 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: cephadm 2026-03-06T21:20:27.099981+0000 mgr.vm03.uwuzgl (mgr.14199) 180 : cephadm [INF] Deploying daemon haproxy.nfs.happy.vm08.kqaecw on vm08 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: cephadm 2026-03-06T21:20:27.099981+0000 mgr.vm03.uwuzgl (mgr.14199) 180 : cephadm [INF] Deploying daemon haproxy.nfs.happy.vm08.kqaecw on vm08 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:27.558833+0000 mon.vm03 (mon.0) 735 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:28.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:27 vm08 bash[20828]: audit 2026-03-06T21:20:27.558833+0000 mon.vm03 (mon.0) 735 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:29.128 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:28 vm08 bash[20828]: cluster 2026-03-06T21:20:28.087855+0000 mgr.vm03.uwuzgl (mgr.14199) 181 : cluster [DBG] pgmap v121: 97 pgs: 97 active+clean; 451 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 3.3 KiB/s wr, 9 op/s 2026-03-06T22:20:29.128 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:28 vm08 bash[20828]: cluster 2026-03-06T21:20:28.087855+0000 mgr.vm03.uwuzgl (mgr.14199) 181 : cluster [DBG] pgmap v121: 97 pgs: 97 active+clean; 451 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 3.3 KiB/s wr, 9 op/s 2026-03-06T22:20:29.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:28 vm03 bash[17055]: cluster 2026-03-06T21:20:28.087855+0000 mgr.vm03.uwuzgl (mgr.14199) 181 : cluster [DBG] pgmap v121: 97 pgs: 97 active+clean; 451 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 3.3 KiB/s wr, 9 op/s 2026-03-06T22:20:29.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:28 vm03 bash[17055]: cluster 2026-03-06T21:20:28.087855+0000 mgr.vm03.uwuzgl (mgr.14199) 181 : cluster [DBG] pgmap v121: 97 pgs: 97 active+clean; 451 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 3.3 KiB/s wr, 9 op/s 2026-03-06T22:20:30.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:29 vm03 bash[17055]: cluster 2026-03-06T21:20:28.878458+0000 mon.vm03 (mon.0) 736 : cluster [DBG] mgrmap e21: vm03.uwuzgl(active, since 3m), standbys: vm08.tdhano 2026-03-06T22:20:30.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:29 vm03 bash[17055]: cluster 2026-03-06T21:20:28.878458+0000 mon.vm03 (mon.0) 736 : cluster [DBG] mgrmap e21: vm03.uwuzgl(active, since 3m), standbys: vm08.tdhano 2026-03-06T22:20:30.209 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:29 vm08 bash[20828]: cluster 2026-03-06T21:20:28.878458+0000 mon.vm03 (mon.0) 736 : cluster [DBG] mgrmap e21: vm03.uwuzgl(active, since 3m), standbys: vm08.tdhano 2026-03-06T22:20:30.209 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:29 vm08 bash[20828]: cluster 2026-03-06T21:20:28.878458+0000 mon.vm03 (mon.0) 736 : cluster [DBG] mgrmap e21: vm03.uwuzgl(active, since 3m), standbys: vm08.tdhano 2026-03-06T22:20:31.327 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:31 vm08 bash[20828]: cluster 2026-03-06T21:20:30.088199+0000 mgr.vm03.uwuzgl (mgr.14199) 182 : cluster [DBG] pgmap v122: 97 pgs: 97 active+clean; 452 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.4 KiB/s rd, 3.1 KiB/s wr, 8 op/s 2026-03-06T22:20:31.327 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:31 vm08 bash[20828]: cluster 2026-03-06T21:20:30.088199+0000 mgr.vm03.uwuzgl (mgr.14199) 182 : cluster [DBG] pgmap v122: 97 pgs: 97 active+clean; 452 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.4 KiB/s rd, 3.1 KiB/s wr, 8 op/s 2026-03-06T22:20:31.327 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:31 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:20:31.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:31 vm03 bash[17055]: cluster 2026-03-06T21:20:30.088199+0000 mgr.vm03.uwuzgl (mgr.14199) 182 : cluster [DBG] pgmap v122: 97 pgs: 97 active+clean; 452 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.4 KiB/s rd, 3.1 KiB/s wr, 8 op/s 2026-03-06T22:20:31.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:31 vm03 bash[17055]: cluster 2026-03-06T21:20:30.088199+0000 mgr.vm03.uwuzgl (mgr.14199) 182 : cluster [DBG] pgmap v122: 97 pgs: 97 active+clean; 452 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.4 KiB/s rd, 3.1 KiB/s wr, 8 op/s 2026-03-06T22:20:31.605 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:31 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:20:31.821 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:20:32.199 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:20:32.199 INFO:teuthology.orchestra.run.vm03.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-06T21:16:05.766418Z", "last_refresh": "2026-03-06T21:20:26.091174Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-06T21:17:13.981567Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-06T21:16:03.845298Z", "last_refresh": "2026-03-06T21:20:25.836604Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:14.763843Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-06T21:16:03.393952Z", "last_refresh": "2026-03-06T21:20:25.836548Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-06T21:16:04.987741Z", "last_refresh": "2026-03-06T21:20:26.090946Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-06T21:20:31.563393Z service:ingress.nfs.happy [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "nfs.happy", "service_name": "ingress.nfs.happy", "service_type": "ingress", "spec": {"backend_service": "nfs.happy", "enable_haproxy_protocol": true, "first_virtual_router_id": 50, "frontend_port": 2049, "monitor_port": 9049, "virtual_ip": "12.12.1.103"}, "status": {"created": "2026-03-06T21:20:21.847175Z", "ports": [2049, 9049], "running": 0, "size": 4, "virtual_ip": "12.12.1.103"}}, {"events": ["2026-03-06T21:20:20.634732Z service:mds.fs1 [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "fs1", "service_name": "mds.fs1", "service_type": "mds", "status": {"created": "2026-03-06T21:20:14.064886Z", "last_refresh": "2026-03-06T21:20:25.836631Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:16.189007Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T21:16:02.939076Z", "last_refresh": "2026-03-06T21:20:25.836298Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:17.196289Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm03:192.168.123.103=vm03", "vm08:192.168.123.108=vm08"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T21:17:07.022646Z", "last_refresh": "2026-03-06T21:20:25.836404Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:20:27.089986Z service:nfs.happy [INFO] \"service was created\""], "placement": {"count": 1}, "service_id": "happy", "service_name": "nfs.happy", "service_type": "nfs", "spec": {"enable_haproxy_protocol": true, "port": 12049}, "status": {"created": "2026-03-06T21:20:21.842604Z", "ports": [12049], "running": 0, "size": 1}}, {"events": ["2026-03-06T21:17:15.412676Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-06T21:16:05.375475Z", "last_refresh": "2026-03-06T21:20:25.836509Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:35.462286Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-06T21:17:35.457628Z", "last_refresh": "2026-03-06T21:20:25.836347Z", "running": 8, "size": 8}}, {"events": ["2026-03-06T21:17:17.200217Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-06T21:16:04.593189Z", "last_refresh": "2026-03-06T21:20:26.091257Z", "ports": [9095], "running": 1, "size": 1}}] 2026-03-06T22:20:32.269 INFO:tasks.cephadm:nfs.happy has 0/1 2026-03-06T22:20:32.842 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:32 vm03 bash[17055]: audit 2026-03-06T21:20:31.554579+0000 mon.vm03 (mon.0) 737 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:32.842 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:32 vm03 bash[17055]: audit 2026-03-06T21:20:31.554579+0000 mon.vm03 (mon.0) 737 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:32.842 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:32 vm03 bash[17055]: audit 2026-03-06T21:20:31.559082+0000 mon.vm03 (mon.0) 738 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:32.842 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:32 vm03 bash[17055]: audit 2026-03-06T21:20:31.559082+0000 mon.vm03 (mon.0) 738 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:32.842 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:32 vm03 bash[17055]: audit 2026-03-06T21:20:31.563217+0000 mon.vm03 (mon.0) 739 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:32.842 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:32 vm03 bash[17055]: audit 2026-03-06T21:20:31.563217+0000 mon.vm03 (mon.0) 739 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:32.842 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:32 vm03 bash[17055]: cephadm 2026-03-06T21:20:31.564407+0000 mgr.vm03.uwuzgl (mgr.14199) 183 : cephadm [INF] Deploying daemon haproxy.nfs.happy.vm03.agxegy on vm03 2026-03-06T22:20:32.842 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:32 vm03 bash[17055]: cephadm 2026-03-06T21:20:31.564407+0000 mgr.vm03.uwuzgl (mgr.14199) 183 : cephadm [INF] Deploying daemon haproxy.nfs.happy.vm03.agxegy on vm03 2026-03-06T22:20:33.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:32 vm08 bash[20828]: audit 2026-03-06T21:20:31.554579+0000 mon.vm03 (mon.0) 737 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:33.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:32 vm08 bash[20828]: audit 2026-03-06T21:20:31.554579+0000 mon.vm03 (mon.0) 737 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:33.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:32 vm08 bash[20828]: audit 2026-03-06T21:20:31.559082+0000 mon.vm03 (mon.0) 738 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:33.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:32 vm08 bash[20828]: audit 2026-03-06T21:20:31.559082+0000 mon.vm03 (mon.0) 738 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:33.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:32 vm08 bash[20828]: audit 2026-03-06T21:20:31.563217+0000 mon.vm03 (mon.0) 739 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:33.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:32 vm08 bash[20828]: audit 2026-03-06T21:20:31.563217+0000 mon.vm03 (mon.0) 739 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:33.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:32 vm08 bash[20828]: cephadm 2026-03-06T21:20:31.564407+0000 mgr.vm03.uwuzgl (mgr.14199) 183 : cephadm [INF] Deploying daemon haproxy.nfs.happy.vm03.agxegy on vm03 2026-03-06T22:20:33.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:32 vm08 bash[20828]: cephadm 2026-03-06T21:20:31.564407+0000 mgr.vm03.uwuzgl (mgr.14199) 183 : cephadm [INF] Deploying daemon haproxy.nfs.happy.vm03.agxegy on vm03 2026-03-06T22:20:33.270 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph orch ls -f json 2026-03-06T22:20:33.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:33 vm03 bash[17055]: cluster 2026-03-06T21:20:32.088535+0000 mgr.vm03.uwuzgl (mgr.14199) 184 : cluster [DBG] pgmap v123: 97 pgs: 97 active+clean; 452 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 2.6 KiB/s wr, 7 op/s 2026-03-06T22:20:33.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:33 vm03 bash[17055]: cluster 2026-03-06T21:20:32.088535+0000 mgr.vm03.uwuzgl (mgr.14199) 184 : cluster [DBG] pgmap v123: 97 pgs: 97 active+clean; 452 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 2.6 KiB/s wr, 7 op/s 2026-03-06T22:20:33.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:33 vm03 bash[17055]: audit 2026-03-06T21:20:32.191299+0000 mgr.vm03.uwuzgl (mgr.14199) 185 : audit [DBG] from='client.14518 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:33.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:33 vm03 bash[17055]: audit 2026-03-06T21:20:32.191299+0000 mgr.vm03.uwuzgl (mgr.14199) 185 : audit [DBG] from='client.14518 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:33 vm08 bash[20828]: cluster 2026-03-06T21:20:32.088535+0000 mgr.vm03.uwuzgl (mgr.14199) 184 : cluster [DBG] pgmap v123: 97 pgs: 97 active+clean; 452 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 2.6 KiB/s wr, 7 op/s 2026-03-06T22:20:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:33 vm08 bash[20828]: cluster 2026-03-06T21:20:32.088535+0000 mgr.vm03.uwuzgl (mgr.14199) 184 : cluster [DBG] pgmap v123: 97 pgs: 97 active+clean; 452 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 2.6 KiB/s wr, 7 op/s 2026-03-06T22:20:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:33 vm08 bash[20828]: audit 2026-03-06T21:20:32.191299+0000 mgr.vm03.uwuzgl (mgr.14199) 185 : audit [DBG] from='client.14518 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:34.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:33 vm08 bash[20828]: audit 2026-03-06T21:20:32.191299+0000 mgr.vm03.uwuzgl (mgr.14199) 185 : audit [DBG] from='client.14518 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:34.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:34 vm03 bash[17055]: cluster 2026-03-06T21:20:34.088951+0000 mgr.vm03.uwuzgl (mgr.14199) 186 : cluster [DBG] pgmap v124: 97 pgs: 97 active+clean; 452 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 2.1 KiB/s wr, 5 op/s 2026-03-06T22:20:34.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:34 vm03 bash[17055]: cluster 2026-03-06T21:20:34.088951+0000 mgr.vm03.uwuzgl (mgr.14199) 186 : cluster [DBG] pgmap v124: 97 pgs: 97 active+clean; 452 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 2.1 KiB/s wr, 5 op/s 2026-03-06T22:20:35.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:34 vm08 bash[20828]: cluster 2026-03-06T21:20:34.088951+0000 mgr.vm03.uwuzgl (mgr.14199) 186 : cluster [DBG] pgmap v124: 97 pgs: 97 active+clean; 452 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 2.1 KiB/s wr, 5 op/s 2026-03-06T22:20:35.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:34 vm08 bash[20828]: cluster 2026-03-06T21:20:34.088951+0000 mgr.vm03.uwuzgl (mgr.14199) 186 : cluster [DBG] pgmap v124: 97 pgs: 97 active+clean; 452 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 2.1 KiB/s wr, 5 op/s 2026-03-06T22:20:36.135 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:35 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:20:36.136 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:36 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:20:36.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:36 vm03 bash[17055]: audit 2026-03-06T21:20:36.059118+0000 mon.vm03 (mon.0) 740 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:20:36.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:36 vm03 bash[17055]: audit 2026-03-06T21:20:36.059118+0000 mon.vm03 (mon.0) 740 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:20:36.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:36 vm08 bash[20828]: audit 2026-03-06T21:20:36.059118+0000 mon.vm03 (mon.0) 740 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:20:36.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:36 vm08 bash[20828]: audit 2026-03-06T21:20:36.059118+0000 mon.vm03 (mon.0) 740 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:20:37.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:37 vm08 bash[20828]: cluster 2026-03-06T21:20:36.089318+0000 mgr.vm03.uwuzgl (mgr.14199) 187 : cluster [DBG] pgmap v125: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1.6 KiB/s wr, 4 op/s 2026-03-06T22:20:37.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:37 vm08 bash[20828]: cluster 2026-03-06T21:20:36.089318+0000 mgr.vm03.uwuzgl (mgr.14199) 187 : cluster [DBG] pgmap v125: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1.6 KiB/s wr, 4 op/s 2026-03-06T22:20:37.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:37 vm08 bash[20828]: audit 2026-03-06T21:20:36.207157+0000 mon.vm03 (mon.0) 741 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:37.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:37 vm08 bash[20828]: audit 2026-03-06T21:20:36.207157+0000 mon.vm03 (mon.0) 741 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:37.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:37 vm08 bash[20828]: audit 2026-03-06T21:20:36.211719+0000 mon.vm03 (mon.0) 742 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:37.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:37 vm08 bash[20828]: audit 2026-03-06T21:20:36.211719+0000 mon.vm03 (mon.0) 742 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:37.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:37 vm08 bash[20828]: audit 2026-03-06T21:20:36.216116+0000 mon.vm03 (mon.0) 743 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:37.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:37 vm08 bash[20828]: audit 2026-03-06T21:20:36.216116+0000 mon.vm03 (mon.0) 743 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:37.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:37 vm08 bash[20828]: audit 2026-03-06T21:20:36.221451+0000 mon.vm03 (mon.0) 744 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:37.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:37 vm08 bash[20828]: audit 2026-03-06T21:20:36.221451+0000 mon.vm03 (mon.0) 744 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:37.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:37 vm08 bash[20828]: cephadm 2026-03-06T21:20:36.221895+0000 mgr.vm03.uwuzgl (mgr.14199) 188 : cephadm [INF] 12.12.1.103 is in 12.12.0.0/22 on vm03 interface ens3 2026-03-06T22:20:37.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:37 vm08 bash[20828]: cephadm 2026-03-06T21:20:36.221895+0000 mgr.vm03.uwuzgl (mgr.14199) 188 : cephadm [INF] 12.12.1.103 is in 12.12.0.0/22 on vm03 interface ens3 2026-03-06T22:20:37.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:37 vm08 bash[20828]: cephadm 2026-03-06T21:20:36.221946+0000 mgr.vm03.uwuzgl (mgr.14199) 189 : cephadm [INF] 12.12.1.103 is in 12.12.0.0/22 on vm08 interface ens3 2026-03-06T22:20:37.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:37 vm08 bash[20828]: cephadm 2026-03-06T21:20:36.221946+0000 mgr.vm03.uwuzgl (mgr.14199) 189 : cephadm [INF] 12.12.1.103 is in 12.12.0.0/22 on vm08 interface ens3 2026-03-06T22:20:37.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:37 vm08 bash[20828]: cephadm 2026-03-06T21:20:36.224846+0000 mgr.vm03.uwuzgl (mgr.14199) 190 : cephadm [INF] Deploying daemon keepalived.nfs.happy.vm03.vyaroa on vm03 2026-03-06T22:20:37.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:37 vm08 bash[20828]: cephadm 2026-03-06T21:20:36.224846+0000 mgr.vm03.uwuzgl (mgr.14199) 190 : cephadm [INF] Deploying daemon keepalived.nfs.happy.vm03.vyaroa on vm03 2026-03-06T22:20:37.632 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:37 vm03 bash[17055]: cluster 2026-03-06T21:20:36.089318+0000 mgr.vm03.uwuzgl (mgr.14199) 187 : cluster [DBG] pgmap v125: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1.6 KiB/s wr, 4 op/s 2026-03-06T22:20:37.632 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:37 vm03 bash[17055]: cluster 2026-03-06T21:20:36.089318+0000 mgr.vm03.uwuzgl (mgr.14199) 187 : cluster [DBG] pgmap v125: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1.6 KiB/s wr, 4 op/s 2026-03-06T22:20:37.632 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:37 vm03 bash[17055]: audit 2026-03-06T21:20:36.207157+0000 mon.vm03 (mon.0) 741 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:37.632 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:37 vm03 bash[17055]: audit 2026-03-06T21:20:36.207157+0000 mon.vm03 (mon.0) 741 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:37.632 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:37 vm03 bash[17055]: audit 2026-03-06T21:20:36.211719+0000 mon.vm03 (mon.0) 742 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:37.632 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:37 vm03 bash[17055]: audit 2026-03-06T21:20:36.211719+0000 mon.vm03 (mon.0) 742 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:37.632 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:37 vm03 bash[17055]: audit 2026-03-06T21:20:36.216116+0000 mon.vm03 (mon.0) 743 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:37.632 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:37 vm03 bash[17055]: audit 2026-03-06T21:20:36.216116+0000 mon.vm03 (mon.0) 743 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:37.632 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:37 vm03 bash[17055]: audit 2026-03-06T21:20:36.221451+0000 mon.vm03 (mon.0) 744 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:37.632 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:37 vm03 bash[17055]: audit 2026-03-06T21:20:36.221451+0000 mon.vm03 (mon.0) 744 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:37.632 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:37 vm03 bash[17055]: cephadm 2026-03-06T21:20:36.221895+0000 mgr.vm03.uwuzgl (mgr.14199) 188 : cephadm [INF] 12.12.1.103 is in 12.12.0.0/22 on vm03 interface ens3 2026-03-06T22:20:37.632 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:37 vm03 bash[17055]: cephadm 2026-03-06T21:20:36.221895+0000 mgr.vm03.uwuzgl (mgr.14199) 188 : cephadm [INF] 12.12.1.103 is in 12.12.0.0/22 on vm03 interface ens3 2026-03-06T22:20:37.632 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:37 vm03 bash[17055]: cephadm 2026-03-06T21:20:36.221946+0000 mgr.vm03.uwuzgl (mgr.14199) 189 : cephadm [INF] 12.12.1.103 is in 12.12.0.0/22 on vm08 interface ens3 2026-03-06T22:20:37.632 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:37 vm03 bash[17055]: cephadm 2026-03-06T21:20:36.221946+0000 mgr.vm03.uwuzgl (mgr.14199) 189 : cephadm [INF] 12.12.1.103 is in 12.12.0.0/22 on vm08 interface ens3 2026-03-06T22:20:37.632 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:37 vm03 bash[17055]: cephadm 2026-03-06T21:20:36.224846+0000 mgr.vm03.uwuzgl (mgr.14199) 190 : cephadm [INF] Deploying daemon keepalived.nfs.happy.vm03.vyaroa on vm03 2026-03-06T22:20:37.632 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:37 vm03 bash[17055]: cephadm 2026-03-06T21:20:36.224846+0000 mgr.vm03.uwuzgl (mgr.14199) 190 : cephadm [INF] Deploying daemon keepalived.nfs.happy.vm03.vyaroa on vm03 2026-03-06T22:20:38.054 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:20:38.424 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:20:38.424 INFO:teuthology.orchestra.run.vm03.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-06T21:16:05.766418Z", "last_refresh": "2026-03-06T21:20:26.091174Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-06T21:17:13.981567Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-06T21:16:03.845298Z", "last_refresh": "2026-03-06T21:20:25.836604Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:14.763843Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-06T21:16:03.393952Z", "last_refresh": "2026-03-06T21:20:25.836548Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-06T21:16:04.987741Z", "last_refresh": "2026-03-06T21:20:26.090946Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-06T21:20:36.216368Z service:ingress.nfs.happy [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "nfs.happy", "service_name": "ingress.nfs.happy", "service_type": "ingress", "spec": {"backend_service": "nfs.happy", "enable_haproxy_protocol": true, "first_virtual_router_id": 50, "frontend_port": 2049, "monitor_port": 9049, "virtual_ip": "12.12.1.103"}, "status": {"created": "2026-03-06T21:20:21.847175Z", "ports": [2049, 9049], "running": 0, "size": 4, "virtual_ip": "12.12.1.103"}}, {"events": ["2026-03-06T21:20:20.634732Z service:mds.fs1 [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "fs1", "service_name": "mds.fs1", "service_type": "mds", "status": {"created": "2026-03-06T21:20:14.064886Z", "last_refresh": "2026-03-06T21:20:25.836631Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:16.189007Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T21:16:02.939076Z", "last_refresh": "2026-03-06T21:20:25.836298Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:17.196289Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm03:192.168.123.103=vm03", "vm08:192.168.123.108=vm08"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T21:17:07.022646Z", "last_refresh": "2026-03-06T21:20:25.836404Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:20:27.089986Z service:nfs.happy [INFO] \"service was created\""], "placement": {"count": 1}, "service_id": "happy", "service_name": "nfs.happy", "service_type": "nfs", "spec": {"enable_haproxy_protocol": true, "port": 12049}, "status": {"created": "2026-03-06T21:20:21.842604Z", "ports": [12049], "running": 0, "size": 1}}, {"events": ["2026-03-06T21:17:15.412676Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-06T21:16:05.375475Z", "last_refresh": "2026-03-06T21:20:25.836509Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:35.462286Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-06T21:17:35.457628Z", "last_refresh": "2026-03-06T21:20:25.836347Z", "running": 8, "size": 8}}, {"events": ["2026-03-06T21:17:17.200217Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-06T21:16:04.593189Z", "last_refresh": "2026-03-06T21:20:26.091257Z", "ports": [9095], "running": 1, "size": 1}}] 2026-03-06T22:20:38.496 INFO:tasks.cephadm:nfs.happy has 0/1 2026-03-06T22:20:39.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:39 vm03 bash[17055]: cluster 2026-03-06T21:20:38.089684+0000 mgr.vm03.uwuzgl (mgr.14199) 191 : cluster [DBG] pgmap v126: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1.2 KiB/s wr, 2 op/s 2026-03-06T22:20:39.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:39 vm03 bash[17055]: cluster 2026-03-06T21:20:38.089684+0000 mgr.vm03.uwuzgl (mgr.14199) 191 : cluster [DBG] pgmap v126: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1.2 KiB/s wr, 2 op/s 2026-03-06T22:20:39.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:39 vm03 bash[17055]: audit 2026-03-06T21:20:38.416425+0000 mgr.vm03.uwuzgl (mgr.14199) 192 : audit [DBG] from='client.14522 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:39.441 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:39 vm03 bash[17055]: audit 2026-03-06T21:20:38.416425+0000 mgr.vm03.uwuzgl (mgr.14199) 192 : audit [DBG] from='client.14522 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:39.497 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph orch ls -f json 2026-03-06T22:20:39.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:39 vm08 bash[20828]: cluster 2026-03-06T21:20:38.089684+0000 mgr.vm03.uwuzgl (mgr.14199) 191 : cluster [DBG] pgmap v126: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1.2 KiB/s wr, 2 op/s 2026-03-06T22:20:39.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:39 vm08 bash[20828]: cluster 2026-03-06T21:20:38.089684+0000 mgr.vm03.uwuzgl (mgr.14199) 191 : cluster [DBG] pgmap v126: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1.2 KiB/s wr, 2 op/s 2026-03-06T22:20:39.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:39 vm08 bash[20828]: audit 2026-03-06T21:20:38.416425+0000 mgr.vm03.uwuzgl (mgr.14199) 192 : audit [DBG] from='client.14522 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:39.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:39 vm08 bash[20828]: audit 2026-03-06T21:20:38.416425+0000 mgr.vm03.uwuzgl (mgr.14199) 192 : audit [DBG] from='client.14522 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:41.423 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:41 vm03 bash[17055]: cluster 2026-03-06T21:20:40.090035+0000 mgr.vm03.uwuzgl (mgr.14199) 193 : cluster [DBG] pgmap v127: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 426 B/s wr, 0 op/s 2026-03-06T22:20:41.423 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:41 vm03 bash[17055]: cluster 2026-03-06T21:20:40.090035+0000 mgr.vm03.uwuzgl (mgr.14199) 193 : cluster [DBG] pgmap v127: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 426 B/s wr, 0 op/s 2026-03-06T22:20:41.647 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:41 vm08 bash[20828]: cluster 2026-03-06T21:20:40.090035+0000 mgr.vm03.uwuzgl (mgr.14199) 193 : cluster [DBG] pgmap v127: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 426 B/s wr, 0 op/s 2026-03-06T22:20:41.647 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:41 vm08 bash[20828]: cluster 2026-03-06T21:20:40.090035+0000 mgr.vm03.uwuzgl (mgr.14199) 193 : cluster [DBG] pgmap v127: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 426 B/s wr, 0 op/s 2026-03-06T22:20:41.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:41 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:20:41.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:41 vm03 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:20:42.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:42 vm03 bash[17055]: audit 2026-03-06T21:20:41.764350+0000 mon.vm03 (mon.0) 745 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:42.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:42 vm03 bash[17055]: audit 2026-03-06T21:20:41.764350+0000 mon.vm03 (mon.0) 745 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:42.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:42 vm03 bash[17055]: audit 2026-03-06T21:20:41.775371+0000 mon.vm03 (mon.0) 746 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:42.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:42 vm03 bash[17055]: audit 2026-03-06T21:20:41.775371+0000 mon.vm03 (mon.0) 746 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:42.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:42 vm03 bash[17055]: audit 2026-03-06T21:20:41.787428+0000 mon.vm03 (mon.0) 747 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:42.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:42 vm03 bash[17055]: audit 2026-03-06T21:20:41.787428+0000 mon.vm03 (mon.0) 747 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:42.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:42 vm03 bash[17055]: cephadm 2026-03-06T21:20:41.789768+0000 mgr.vm03.uwuzgl (mgr.14199) 194 : cephadm [INF] 12.12.1.103 is in 12.12.0.0/22 on vm08 interface ens3 2026-03-06T22:20:42.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:42 vm03 bash[17055]: cephadm 2026-03-06T21:20:41.789768+0000 mgr.vm03.uwuzgl (mgr.14199) 194 : cephadm [INF] 12.12.1.103 is in 12.12.0.0/22 on vm08 interface ens3 2026-03-06T22:20:42.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:42 vm03 bash[17055]: cephadm 2026-03-06T21:20:41.789812+0000 mgr.vm03.uwuzgl (mgr.14199) 195 : cephadm [INF] 12.12.1.103 is in 12.12.0.0/22 on vm03 interface ens3 2026-03-06T22:20:42.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:42 vm03 bash[17055]: cephadm 2026-03-06T21:20:41.789812+0000 mgr.vm03.uwuzgl (mgr.14199) 195 : cephadm [INF] 12.12.1.103 is in 12.12.0.0/22 on vm03 interface ens3 2026-03-06T22:20:42.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:42 vm03 bash[17055]: cephadm 2026-03-06T21:20:41.790209+0000 mgr.vm03.uwuzgl (mgr.14199) 196 : cephadm [INF] Deploying daemon keepalived.nfs.happy.vm08.rvikvg on vm08 2026-03-06T22:20:42.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:42 vm03 bash[17055]: cephadm 2026-03-06T21:20:41.790209+0000 mgr.vm03.uwuzgl (mgr.14199) 196 : cephadm [INF] Deploying daemon keepalived.nfs.happy.vm08.rvikvg on vm08 2026-03-06T22:20:42.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:42 vm03 bash[17055]: cluster 2026-03-06T21:20:42.090668+0000 mgr.vm03.uwuzgl (mgr.14199) 197 : cluster [DBG] pgmap v128: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 255 B/s wr, 0 op/s 2026-03-06T22:20:42.941 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:42 vm03 bash[17055]: cluster 2026-03-06T21:20:42.090668+0000 mgr.vm03.uwuzgl (mgr.14199) 197 : cluster [DBG] pgmap v128: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 255 B/s wr, 0 op/s 2026-03-06T22:20:43.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:42 vm08 bash[20828]: audit 2026-03-06T21:20:41.764350+0000 mon.vm03 (mon.0) 745 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:43.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:42 vm08 bash[20828]: audit 2026-03-06T21:20:41.764350+0000 mon.vm03 (mon.0) 745 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:43.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:42 vm08 bash[20828]: audit 2026-03-06T21:20:41.775371+0000 mon.vm03 (mon.0) 746 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:43.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:42 vm08 bash[20828]: audit 2026-03-06T21:20:41.775371+0000 mon.vm03 (mon.0) 746 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:43.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:42 vm08 bash[20828]: audit 2026-03-06T21:20:41.787428+0000 mon.vm03 (mon.0) 747 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:43.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:42 vm08 bash[20828]: audit 2026-03-06T21:20:41.787428+0000 mon.vm03 (mon.0) 747 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:43.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:42 vm08 bash[20828]: cephadm 2026-03-06T21:20:41.789768+0000 mgr.vm03.uwuzgl (mgr.14199) 194 : cephadm [INF] 12.12.1.103 is in 12.12.0.0/22 on vm08 interface ens3 2026-03-06T22:20:43.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:42 vm08 bash[20828]: cephadm 2026-03-06T21:20:41.789768+0000 mgr.vm03.uwuzgl (mgr.14199) 194 : cephadm [INF] 12.12.1.103 is in 12.12.0.0/22 on vm08 interface ens3 2026-03-06T22:20:43.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:42 vm08 bash[20828]: cephadm 2026-03-06T21:20:41.789812+0000 mgr.vm03.uwuzgl (mgr.14199) 195 : cephadm [INF] 12.12.1.103 is in 12.12.0.0/22 on vm03 interface ens3 2026-03-06T22:20:43.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:42 vm08 bash[20828]: cephadm 2026-03-06T21:20:41.789812+0000 mgr.vm03.uwuzgl (mgr.14199) 195 : cephadm [INF] 12.12.1.103 is in 12.12.0.0/22 on vm03 interface ens3 2026-03-06T22:20:43.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:42 vm08 bash[20828]: cephadm 2026-03-06T21:20:41.790209+0000 mgr.vm03.uwuzgl (mgr.14199) 196 : cephadm [INF] Deploying daemon keepalived.nfs.happy.vm08.rvikvg on vm08 2026-03-06T22:20:43.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:42 vm08 bash[20828]: cephadm 2026-03-06T21:20:41.790209+0000 mgr.vm03.uwuzgl (mgr.14199) 196 : cephadm [INF] Deploying daemon keepalived.nfs.happy.vm08.rvikvg on vm08 2026-03-06T22:20:43.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:42 vm08 bash[20828]: cluster 2026-03-06T21:20:42.090668+0000 mgr.vm03.uwuzgl (mgr.14199) 197 : cluster [DBG] pgmap v128: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 255 B/s wr, 0 op/s 2026-03-06T22:20:43.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:42 vm08 bash[20828]: cluster 2026-03-06T21:20:42.090668+0000 mgr.vm03.uwuzgl (mgr.14199) 197 : cluster [DBG] pgmap v128: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 255 B/s wr, 0 op/s 2026-03-06T22:20:44.336 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:20:44.700 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:20:44.700 INFO:teuthology.orchestra.run.vm03.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-06T21:16:05.766418Z", "last_refresh": "2026-03-06T21:20:26.091174Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-06T21:17:13.981567Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-06T21:16:03.845298Z", "last_refresh": "2026-03-06T21:20:25.836604Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:14.763843Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-06T21:16:03.393952Z", "last_refresh": "2026-03-06T21:20:25.836548Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-06T21:16:04.987741Z", "last_refresh": "2026-03-06T21:20:26.090946Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-06T21:20:41.788653Z service:ingress.nfs.happy [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "nfs.happy", "service_name": "ingress.nfs.happy", "service_type": "ingress", "spec": {"backend_service": "nfs.happy", "enable_haproxy_protocol": true, "first_virtual_router_id": 50, "frontend_port": 2049, "monitor_port": 9049, "virtual_ip": "12.12.1.103"}, "status": {"created": "2026-03-06T21:20:21.847175Z", "ports": [2049, 9049], "running": 0, "size": 4, "virtual_ip": "12.12.1.103"}}, {"events": ["2026-03-06T21:20:20.634732Z service:mds.fs1 [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "fs1", "service_name": "mds.fs1", "service_type": "mds", "status": {"created": "2026-03-06T21:20:14.064886Z", "last_refresh": "2026-03-06T21:20:25.836631Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:16.189007Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T21:16:02.939076Z", "last_refresh": "2026-03-06T21:20:25.836298Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:17.196289Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm03:192.168.123.103=vm03", "vm08:192.168.123.108=vm08"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T21:17:07.022646Z", "last_refresh": "2026-03-06T21:20:25.836404Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:20:27.089986Z service:nfs.happy [INFO] \"service was created\""], "placement": {"count": 1}, "service_id": "happy", "service_name": "nfs.happy", "service_type": "nfs", "spec": {"enable_haproxy_protocol": true, "port": 12049}, "status": {"created": "2026-03-06T21:20:21.842604Z", "ports": [12049], "running": 0, "size": 1}}, {"events": ["2026-03-06T21:17:15.412676Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-06T21:16:05.375475Z", "last_refresh": "2026-03-06T21:20:25.836509Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:35.462286Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-06T21:17:35.457628Z", "last_refresh": "2026-03-06T21:20:25.836347Z", "running": 8, "size": 8}}, {"events": ["2026-03-06T21:17:17.200217Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-06T21:16:04.593189Z", "last_refresh": "2026-03-06T21:20:26.091257Z", "ports": [9095], "running": 1, "size": 1}}] 2026-03-06T22:20:44.764 INFO:tasks.cephadm:nfs.happy has 0/1 2026-03-06T22:20:45.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:45 vm08 bash[20828]: cluster 2026-03-06T21:20:44.091130+0000 mgr.vm03.uwuzgl (mgr.14199) 198 : cluster [DBG] pgmap v129: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-06T22:20:45.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:45 vm08 bash[20828]: cluster 2026-03-06T21:20:44.091130+0000 mgr.vm03.uwuzgl (mgr.14199) 198 : cluster [DBG] pgmap v129: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-06T22:20:45.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:45 vm03 bash[17055]: cluster 2026-03-06T21:20:44.091130+0000 mgr.vm03.uwuzgl (mgr.14199) 198 : cluster [DBG] pgmap v129: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-06T22:20:45.440 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:45 vm03 bash[17055]: cluster 2026-03-06T21:20:44.091130+0000 mgr.vm03.uwuzgl (mgr.14199) 198 : cluster [DBG] pgmap v129: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-06T22:20:45.765 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph orch ls -f json 2026-03-06T22:20:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:46 vm08 bash[20828]: audit 2026-03-06T21:20:44.692081+0000 mgr.vm03.uwuzgl (mgr.14199) 199 : audit [DBG] from='client.14526 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:46.344 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:46 vm08 bash[20828]: audit 2026-03-06T21:20:44.692081+0000 mgr.vm03.uwuzgl (mgr.14199) 199 : audit [DBG] from='client.14526 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:46.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:46 vm03 bash[17055]: audit 2026-03-06T21:20:44.692081+0000 mgr.vm03.uwuzgl (mgr.14199) 199 : audit [DBG] from='client.14526 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:46.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:46 vm03 bash[17055]: audit 2026-03-06T21:20:44.692081+0000 mgr.vm03.uwuzgl (mgr.14199) 199 : audit [DBG] from='client.14526 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:46.967 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:46 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:20:46.967 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:46 vm08 systemd[1]: /etc/systemd/system/ceph-894e000c-19a1-11f1-8dbe-23b24380a082@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-06T22:20:47.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:47 vm08 bash[20828]: cluster 2026-03-06T21:20:46.091514+0000 mgr.vm03.uwuzgl (mgr.14199) 200 : cluster [DBG] pgmap v130: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 255 B/s wr, 0 op/s 2026-03-06T22:20:47.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:47 vm08 bash[20828]: cluster 2026-03-06T21:20:46.091514+0000 mgr.vm03.uwuzgl (mgr.14199) 200 : cluster [DBG] pgmap v130: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 255 B/s wr, 0 op/s 2026-03-06T22:20:47.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:47 vm08 bash[20828]: audit 2026-03-06T21:20:47.115564+0000 mon.vm03 (mon.0) 748 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:47.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:47 vm08 bash[20828]: audit 2026-03-06T21:20:47.115564+0000 mon.vm03 (mon.0) 748 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:47.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:47 vm08 bash[20828]: audit 2026-03-06T21:20:47.122807+0000 mon.vm03 (mon.0) 749 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:47.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:47 vm08 bash[20828]: audit 2026-03-06T21:20:47.122807+0000 mon.vm03 (mon.0) 749 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:47.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:47 vm08 bash[20828]: audit 2026-03-06T21:20:47.127098+0000 mon.vm03 (mon.0) 750 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:47.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:47 vm08 bash[20828]: audit 2026-03-06T21:20:47.127098+0000 mon.vm03 (mon.0) 750 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:47.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:47 vm08 bash[20828]: audit 2026-03-06T21:20:47.131222+0000 mon.vm03 (mon.0) 751 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:47.501 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:47 vm08 bash[20828]: audit 2026-03-06T21:20:47.131222+0000 mon.vm03 (mon.0) 751 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:47.502 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:47 vm08 bash[20828]: audit 2026-03-06T21:20:47.146057+0000 mon.vm03 (mon.0) 752 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:20:47.502 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:47 vm08 bash[20828]: audit 2026-03-06T21:20:47.146057+0000 mon.vm03 (mon.0) 752 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:20:47.690 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:47 vm03 bash[17055]: cluster 2026-03-06T21:20:46.091514+0000 mgr.vm03.uwuzgl (mgr.14199) 200 : cluster [DBG] pgmap v130: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 255 B/s wr, 0 op/s 2026-03-06T22:20:47.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:47 vm03 bash[17055]: cluster 2026-03-06T21:20:46.091514+0000 mgr.vm03.uwuzgl (mgr.14199) 200 : cluster [DBG] pgmap v130: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 255 B/s wr, 0 op/s 2026-03-06T22:20:47.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:47 vm03 bash[17055]: audit 2026-03-06T21:20:47.115564+0000 mon.vm03 (mon.0) 748 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:47.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:47 vm03 bash[17055]: audit 2026-03-06T21:20:47.115564+0000 mon.vm03 (mon.0) 748 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:47.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:47 vm03 bash[17055]: audit 2026-03-06T21:20:47.122807+0000 mon.vm03 (mon.0) 749 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:47.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:47 vm03 bash[17055]: audit 2026-03-06T21:20:47.122807+0000 mon.vm03 (mon.0) 749 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:47.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:47 vm03 bash[17055]: audit 2026-03-06T21:20:47.127098+0000 mon.vm03 (mon.0) 750 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:47.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:47 vm03 bash[17055]: audit 2026-03-06T21:20:47.127098+0000 mon.vm03 (mon.0) 750 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:47.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:47 vm03 bash[17055]: audit 2026-03-06T21:20:47.131222+0000 mon.vm03 (mon.0) 751 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:47.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:47 vm03 bash[17055]: audit 2026-03-06T21:20:47.131222+0000 mon.vm03 (mon.0) 751 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:47.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:47 vm03 bash[17055]: audit 2026-03-06T21:20:47.146057+0000 mon.vm03 (mon.0) 752 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:20:47.691 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:47 vm03 bash[17055]: audit 2026-03-06T21:20:47.146057+0000 mon.vm03 (mon.0) 752 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:20:48.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:48 vm08 bash[20828]: audit 2026-03-06T21:20:47.567416+0000 mon.vm03 (mon.0) 753 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:48.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:48 vm08 bash[20828]: audit 2026-03-06T21:20:47.567416+0000 mon.vm03 (mon.0) 753 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:48.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:48 vm03 bash[17055]: audit 2026-03-06T21:20:47.567416+0000 mon.vm03 (mon.0) 753 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:48.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:48 vm03 bash[17055]: audit 2026-03-06T21:20:47.567416+0000 mon.vm03 (mon.0) 753 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:50.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:49 vm08 bash[20828]: cluster 2026-03-06T21:20:48.092036+0000 mgr.vm03.uwuzgl (mgr.14199) 201 : cluster [DBG] pgmap v131: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:50.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:49 vm08 bash[20828]: cluster 2026-03-06T21:20:48.092036+0000 mgr.vm03.uwuzgl (mgr.14199) 201 : cluster [DBG] pgmap v131: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:50.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:49 vm03 bash[17055]: cluster 2026-03-06T21:20:48.092036+0000 mgr.vm03.uwuzgl (mgr.14199) 201 : cluster [DBG] pgmap v131: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:50.190 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:49 vm03 bash[17055]: cluster 2026-03-06T21:20:48.092036+0000 mgr.vm03.uwuzgl (mgr.14199) 201 : cluster [DBG] pgmap v131: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:50.608 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:20:50.939 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:50 vm03 bash[17055]: cluster 2026-03-06T21:20:50.092393+0000 mgr.vm03.uwuzgl (mgr.14199) 202 : cluster [DBG] pgmap v132: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:50.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:50 vm03 bash[17055]: cluster 2026-03-06T21:20:50.092393+0000 mgr.vm03.uwuzgl (mgr.14199) 202 : cluster [DBG] pgmap v132: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:50.974 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:20:50.974 INFO:teuthology.orchestra.run.vm03.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-06T21:16:05.766418Z", "last_refresh": "2026-03-06T21:20:26.091174Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-06T21:17:13.981567Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-06T21:16:03.845298Z", "last_refresh": "2026-03-06T21:20:25.836604Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:14.763843Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-06T21:16:03.393952Z", "last_refresh": "2026-03-06T21:20:25.836548Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-06T21:16:04.987741Z", "last_refresh": "2026-03-06T21:20:26.090946Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-06T21:20:47.131435Z service:ingress.nfs.happy [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "nfs.happy", "service_name": "ingress.nfs.happy", "service_type": "ingress", "spec": {"backend_service": "nfs.happy", "enable_haproxy_protocol": true, "first_virtual_router_id": 50, "frontend_port": 2049, "monitor_port": 9049, "virtual_ip": "12.12.1.103"}, "status": {"created": "2026-03-06T21:20:21.847175Z", "ports": [2049, 9049], "running": 0, "size": 4, "virtual_ip": "12.12.1.103"}}, {"events": ["2026-03-06T21:20:20.634732Z service:mds.fs1 [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "fs1", "service_name": "mds.fs1", "service_type": "mds", "status": {"created": "2026-03-06T21:20:14.064886Z", "last_refresh": "2026-03-06T21:20:25.836631Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:16.189007Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T21:16:02.939076Z", "last_refresh": "2026-03-06T21:20:25.836298Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:17.196289Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm03:192.168.123.103=vm03", "vm08:192.168.123.108=vm08"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T21:17:07.022646Z", "last_refresh": "2026-03-06T21:20:25.836404Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:20:27.089986Z service:nfs.happy [INFO] \"service was created\""], "placement": {"count": 1}, "service_id": "happy", "service_name": "nfs.happy", "service_type": "nfs", "spec": {"enable_haproxy_protocol": true, "port": 12049}, "status": {"created": "2026-03-06T21:20:21.842604Z", "ports": [12049], "running": 0, "size": 1}}, {"events": ["2026-03-06T21:17:15.412676Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-06T21:16:05.375475Z", "last_refresh": "2026-03-06T21:20:25.836509Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:35.462286Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-06T21:17:35.457628Z", "last_refresh": "2026-03-06T21:20:25.836347Z", "running": 8, "size": 8}}, {"events": ["2026-03-06T21:17:17.200217Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-06T21:16:04.593189Z", "last_refresh": "2026-03-06T21:20:26.091257Z", "ports": [9095], "running": 1, "size": 1}}] 2026-03-06T22:20:51.051 INFO:tasks.cephadm:nfs.happy has 0/1 2026-03-06T22:20:51.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:50 vm08 bash[20828]: cluster 2026-03-06T21:20:50.092393+0000 mgr.vm03.uwuzgl (mgr.14199) 202 : cluster [DBG] pgmap v132: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:51.250 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:50 vm08 bash[20828]: cluster 2026-03-06T21:20:50.092393+0000 mgr.vm03.uwuzgl (mgr.14199) 202 : cluster [DBG] pgmap v132: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:51.939 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:51 vm03 bash[17055]: audit 2026-03-06T21:20:50.966497+0000 mgr.vm03.uwuzgl (mgr.14199) 203 : audit [DBG] from='client.14530 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:51.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:51 vm03 bash[17055]: audit 2026-03-06T21:20:50.966497+0000 mgr.vm03.uwuzgl (mgr.14199) 203 : audit [DBG] from='client.14530 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:51.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:51 vm03 bash[17055]: audit 2026-03-06T21:20:51.059413+0000 mon.vm03 (mon.0) 754 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:20:51.940 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:51 vm03 bash[17055]: audit 2026-03-06T21:20:51.059413+0000 mon.vm03 (mon.0) 754 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:20:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:51 vm08 bash[20828]: audit 2026-03-06T21:20:50.966497+0000 mgr.vm03.uwuzgl (mgr.14199) 203 : audit [DBG] from='client.14530 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:52.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:51 vm08 bash[20828]: audit 2026-03-06T21:20:50.966497+0000 mgr.vm03.uwuzgl (mgr.14199) 203 : audit [DBG] from='client.14530 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:52.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:51 vm08 bash[20828]: audit 2026-03-06T21:20:51.059413+0000 mon.vm03 (mon.0) 754 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:20:52.001 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:51 vm08 bash[20828]: audit 2026-03-06T21:20:51.059413+0000 mon.vm03 (mon.0) 754 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:20:52.052 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph orch ls -f json 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: cluster 2026-03-06T21:20:52.092785+0000 mgr.vm03.uwuzgl (mgr.14199) 204 : cluster [DBG] pgmap v133: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: cluster 2026-03-06T21:20:52.092785+0000 mgr.vm03.uwuzgl (mgr.14199) 204 : cluster [DBG] pgmap v133: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.406565+0000 mon.vm03 (mon.0) 755 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.406565+0000 mon.vm03 (mon.0) 755 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.418364+0000 mon.vm03 (mon.0) 756 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.418364+0000 mon.vm03 (mon.0) 756 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.461146+0000 mon.vm03 (mon.0) 757 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.461146+0000 mon.vm03 (mon.0) 757 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.466700+0000 mon.vm03 (mon.0) 758 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.466700+0000 mon.vm03 (mon.0) 758 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.803495+0000 mon.vm03 (mon.0) 759 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.803495+0000 mon.vm03 (mon.0) 759 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.804010+0000 mon.vm03 (mon.0) 760 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.804010+0000 mon.vm03 (mon.0) 760 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.808529+0000 mon.vm03 (mon.0) 761 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.808529+0000 mon.vm03 (mon.0) 761 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.810511+0000 mon.vm03 (mon.0) 762 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.810511+0000 mon.vm03 (mon.0) 762 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.831216+0000 mon.vm03 (mon.0) 763 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.589 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:53 vm03 bash[17055]: audit 2026-03-06T21:20:52.831216+0000 mon.vm03 (mon.0) 763 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: cluster 2026-03-06T21:20:52.092785+0000 mgr.vm03.uwuzgl (mgr.14199) 204 : cluster [DBG] pgmap v133: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:53.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: cluster 2026-03-06T21:20:52.092785+0000 mgr.vm03.uwuzgl (mgr.14199) 204 : cluster [DBG] pgmap v133: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:53.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.406565+0000 mon.vm03 (mon.0) 755 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.406565+0000 mon.vm03 (mon.0) 755 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.418364+0000 mon.vm03 (mon.0) 756 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.418364+0000 mon.vm03 (mon.0) 756 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.461146+0000 mon.vm03 (mon.0) 757 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.461146+0000 mon.vm03 (mon.0) 757 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.466700+0000 mon.vm03 (mon.0) 758 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.466700+0000 mon.vm03 (mon.0) 758 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.803495+0000 mon.vm03 (mon.0) 759 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:53.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.803495+0000 mon.vm03 (mon.0) 759 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:53.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.804010+0000 mon.vm03 (mon.0) 760 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:20:53.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.804010+0000 mon.vm03 (mon.0) 760 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:20:53.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.808529+0000 mon.vm03 (mon.0) 761 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.808529+0000 mon.vm03 (mon.0) 761 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.810511+0000 mon.vm03 (mon.0) 762 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:20:53.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.810511+0000 mon.vm03 (mon.0) 762 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:20:53.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.831216+0000 mon.vm03 (mon.0) 763 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:53.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:53 vm08 bash[20828]: audit 2026-03-06T21:20:52.831216+0000 mon.vm03 (mon.0) 763 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:54.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:54 vm03 bash[17055]: cephadm 2026-03-06T21:20:52.842487+0000 mgr.vm03.uwuzgl (mgr.14199) 205 : cephadm [INF] Reconfiguring prometheus.vm03 (dependencies changed)... 2026-03-06T22:20:54.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:54 vm03 bash[17055]: cephadm 2026-03-06T21:20:52.842487+0000 mgr.vm03.uwuzgl (mgr.14199) 205 : cephadm [INF] Reconfiguring prometheus.vm03 (dependencies changed)... 2026-03-06T22:20:54.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:54 vm03 bash[17055]: cephadm 2026-03-06T21:20:53.005226+0000 mgr.vm03.uwuzgl (mgr.14199) 206 : cephadm [INF] Reconfiguring daemon prometheus.vm03 on vm03 2026-03-06T22:20:54.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:54 vm03 bash[17055]: cephadm 2026-03-06T21:20:53.005226+0000 mgr.vm03.uwuzgl (mgr.14199) 206 : cephadm [INF] Reconfiguring daemon prometheus.vm03 on vm03 2026-03-06T22:20:54.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:54 vm03 bash[17055]: audit 2026-03-06T21:20:53.667562+0000 mon.vm03 (mon.0) 764 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:54.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:54 vm03 bash[17055]: audit 2026-03-06T21:20:53.667562+0000 mon.vm03 (mon.0) 764 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:54.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:54 vm03 bash[17055]: audit 2026-03-06T21:20:53.673907+0000 mon.vm03 (mon.0) 765 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:54.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:54 vm03 bash[17055]: audit 2026-03-06T21:20:53.673907+0000 mon.vm03 (mon.0) 765 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:54.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:54 vm03 bash[17055]: audit 2026-03-06T21:20:53.679638+0000 mon.vm03 (mon.0) 766 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T22:20:54.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:54 vm03 bash[17055]: audit 2026-03-06T21:20:53.679638+0000 mon.vm03 (mon.0) 766 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T22:20:54.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:54 vm03 bash[17055]: audit 2026-03-06T21:20:53.723967+0000 mon.vm03 (mon.0) 767 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:20:54.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:54 vm03 bash[17055]: audit 2026-03-06T21:20:53.723967+0000 mon.vm03 (mon.0) 767 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:20:54.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:54 vm08 bash[20828]: cephadm 2026-03-06T21:20:52.842487+0000 mgr.vm03.uwuzgl (mgr.14199) 205 : cephadm [INF] Reconfiguring prometheus.vm03 (dependencies changed)... 2026-03-06T22:20:54.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:54 vm08 bash[20828]: cephadm 2026-03-06T21:20:52.842487+0000 mgr.vm03.uwuzgl (mgr.14199) 205 : cephadm [INF] Reconfiguring prometheus.vm03 (dependencies changed)... 2026-03-06T22:20:54.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:54 vm08 bash[20828]: cephadm 2026-03-06T21:20:53.005226+0000 mgr.vm03.uwuzgl (mgr.14199) 206 : cephadm [INF] Reconfiguring daemon prometheus.vm03 on vm03 2026-03-06T22:20:54.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:54 vm08 bash[20828]: cephadm 2026-03-06T21:20:53.005226+0000 mgr.vm03.uwuzgl (mgr.14199) 206 : cephadm [INF] Reconfiguring daemon prometheus.vm03 on vm03 2026-03-06T22:20:54.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:54 vm08 bash[20828]: audit 2026-03-06T21:20:53.667562+0000 mon.vm03 (mon.0) 764 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:54.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:54 vm08 bash[20828]: audit 2026-03-06T21:20:53.667562+0000 mon.vm03 (mon.0) 764 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:54.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:54 vm08 bash[20828]: audit 2026-03-06T21:20:53.673907+0000 mon.vm03 (mon.0) 765 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:54.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:54 vm08 bash[20828]: audit 2026-03-06T21:20:53.673907+0000 mon.vm03 (mon.0) 765 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:54.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:54 vm08 bash[20828]: audit 2026-03-06T21:20:53.679638+0000 mon.vm03 (mon.0) 766 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T22:20:54.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:54 vm08 bash[20828]: audit 2026-03-06T21:20:53.679638+0000 mon.vm03 (mon.0) 766 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T22:20:54.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:54 vm08 bash[20828]: audit 2026-03-06T21:20:53.723967+0000 mon.vm03 (mon.0) 767 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:20:54.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:54 vm08 bash[20828]: audit 2026-03-06T21:20:53.723967+0000 mon.vm03 (mon.0) 767 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T22:20:55.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:55 vm03 bash[17055]: audit 2026-03-06T21:20:53.679920+0000 mgr.vm03.uwuzgl (mgr.14199) 207 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T22:20:55.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:55 vm03 bash[17055]: audit 2026-03-06T21:20:53.679920+0000 mgr.vm03.uwuzgl (mgr.14199) 207 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T22:20:55.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:55 vm03 bash[17055]: cluster 2026-03-06T21:20:54.093193+0000 mgr.vm03.uwuzgl (mgr.14199) 208 : cluster [DBG] pgmap v134: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 0 op/s 2026-03-06T22:20:55.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:55 vm03 bash[17055]: cluster 2026-03-06T21:20:54.093193+0000 mgr.vm03.uwuzgl (mgr.14199) 208 : cluster [DBG] pgmap v134: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 0 op/s 2026-03-06T22:20:55.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:55 vm08 bash[20828]: audit 2026-03-06T21:20:53.679920+0000 mgr.vm03.uwuzgl (mgr.14199) 207 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T22:20:55.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:55 vm08 bash[20828]: audit 2026-03-06T21:20:53.679920+0000 mgr.vm03.uwuzgl (mgr.14199) 207 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-06T22:20:55.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:55 vm08 bash[20828]: cluster 2026-03-06T21:20:54.093193+0000 mgr.vm03.uwuzgl (mgr.14199) 208 : cluster [DBG] pgmap v134: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 0 op/s 2026-03-06T22:20:55.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:55 vm08 bash[20828]: cluster 2026-03-06T21:20:54.093193+0000 mgr.vm03.uwuzgl (mgr.14199) 208 : cluster [DBG] pgmap v134: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 0 op/s 2026-03-06T22:20:57.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:57 vm03 bash[17055]: cluster 2026-03-06T21:20:56.093490+0000 mgr.vm03.uwuzgl (mgr.14199) 209 : cluster [DBG] pgmap v135: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:57.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:57 vm03 bash[17055]: cluster 2026-03-06T21:20:56.093490+0000 mgr.vm03.uwuzgl (mgr.14199) 209 : cluster [DBG] pgmap v135: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:57.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:57 vm08 bash[20828]: cluster 2026-03-06T21:20:56.093490+0000 mgr.vm03.uwuzgl (mgr.14199) 209 : cluster [DBG] pgmap v135: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:57.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:57 vm08 bash[20828]: cluster 2026-03-06T21:20:56.093490+0000 mgr.vm03.uwuzgl (mgr.14199) 209 : cluster [DBG] pgmap v135: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:57.969 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:20:58.396 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:20:58.396 INFO:teuthology.orchestra.run.vm03.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-06T21:16:05.766418Z", "last_refresh": "2026-03-06T21:20:52.455533Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-06T21:17:13.981567Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-06T21:16:03.845298Z", "last_refresh": "2026-03-06T21:20:52.400696Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:14.763843Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-06T21:16:03.393952Z", "last_refresh": "2026-03-06T21:20:52.400641Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-06T21:16:04.987741Z", "last_refresh": "2026-03-06T21:20:52.455263Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-06T21:20:47.131435Z service:ingress.nfs.happy [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "nfs.happy", "service_name": "ingress.nfs.happy", "service_type": "ingress", "spec": {"backend_service": "nfs.happy", "enable_haproxy_protocol": true, "first_virtual_router_id": 50, "frontend_port": 2049, "monitor_port": 9049, "virtual_ip": "12.12.1.103"}, "status": {"created": "2026-03-06T21:20:21.847175Z", "last_refresh": "2026-03-06T21:20:52.400585Z", "ports": [2049, 9049], "running": 4, "size": 4, "virtual_ip": "12.12.1.103"}}, {"events": ["2026-03-06T21:20:20.634732Z service:mds.fs1 [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "fs1", "service_name": "mds.fs1", "service_type": "mds", "status": {"created": "2026-03-06T21:20:14.064886Z", "last_refresh": "2026-03-06T21:20:52.400753Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:16.189007Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T21:16:02.939076Z", "last_refresh": "2026-03-06T21:20:52.400330Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:17.196289Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm03:192.168.123.103=vm03", "vm08:192.168.123.108=vm08"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T21:17:07.022646Z", "last_refresh": "2026-03-06T21:20:52.400494Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:20:52.831563Z service:nfs.happy [INFO] \"service was created\""], "placement": {"count": 1}, "service_id": "happy", "service_name": "nfs.happy", "service_type": "nfs", "spec": {"enable_haproxy_protocol": true, "port": 12049}, "status": {"created": "2026-03-06T21:20:21.842604Z", "last_refresh": "2026-03-06T21:20:52.400725Z", "ports": [12049], "running": 1, "size": 1}}, {"events": ["2026-03-06T21:17:15.412676Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-06T21:16:05.375475Z", "last_refresh": "2026-03-06T21:20:52.400557Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:35.462286Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-06T21:17:35.457628Z", "last_refresh": "2026-03-06T21:20:52.400376Z", "running": 8, "size": 8}}, {"events": ["2026-03-06T21:17:17.200217Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-06T21:16:04.593189Z", "ports": [9095], "running": 0, "size": 1}}] 2026-03-06T22:20:58.503 INFO:tasks.cephadm:nfs.happy has 1/1 2026-03-06T22:20:58.503 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-06T22:20:58.505 INFO:tasks.cephadm:Waiting for ceph service ingress.nfs.happy to start (timeout 300)... 2026-03-06T22:20:58.505 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- ceph orch ls -f json 2026-03-06T22:20:59.432 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: cluster 2026-03-06T21:20:58.093845+0000 mgr.vm03.uwuzgl (mgr.14199) 210 : cluster [DBG] pgmap v136: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:59.432 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: cluster 2026-03-06T21:20:58.093845+0000 mgr.vm03.uwuzgl (mgr.14199) 210 : cluster [DBG] pgmap v136: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:59.432 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:58.387231+0000 mgr.vm03.uwuzgl (mgr.14199) 211 : audit [DBG] from='client.14534 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:59.433 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:58.387231+0000 mgr.vm03.uwuzgl (mgr.14199) 211 : audit [DBG] from='client.14534 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: cluster 2026-03-06T21:20:58.093845+0000 mgr.vm03.uwuzgl (mgr.14199) 210 : cluster [DBG] pgmap v136: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: cluster 2026-03-06T21:20:58.093845+0000 mgr.vm03.uwuzgl (mgr.14199) 210 : cluster [DBG] pgmap v136: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:58.387231+0000 mgr.vm03.uwuzgl (mgr.14199) 211 : audit [DBG] from='client.14534 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:58.387231+0000 mgr.vm03.uwuzgl (mgr.14199) 211 : audit [DBG] from='client.14534 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:58.938540+0000 mon.vm03 (mon.0) 768 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:58.938540+0000 mon.vm03 (mon.0) 768 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:58.975339+0000 mon.vm03 (mon.0) 769 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:58.975339+0000 mon.vm03 (mon.0) 769 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:59.028475+0000 mon.vm03 (mon.0) 770 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:59.028475+0000 mon.vm03 (mon.0) 770 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:59.034220+0000 mon.vm03 (mon.0) 771 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:59.034220+0000 mon.vm03 (mon.0) 771 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:59.035196+0000 mon.vm03 (mon.0) 772 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:59.035196+0000 mon.vm03 (mon.0) 772 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:59.035772+0000 mon.vm03 (mon.0) 773 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:59.035772+0000 mon.vm03 (mon.0) 773 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:59.040056+0000 mon.vm03 (mon.0) 774 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:59.040056+0000 mon.vm03 (mon.0) 774 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:59.041578+0000 mon.vm03 (mon.0) 775 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:59.041578+0000 mon.vm03 (mon.0) 775 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:59.048131+0000 mon.vm03 (mon.0) 776 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.689 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:20:59 vm03 bash[17055]: audit 2026-03-06T21:20:59.048131+0000 mon.vm03 (mon.0) 776 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:58.938540+0000 mon.vm03 (mon.0) 768 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:58.938540+0000 mon.vm03 (mon.0) 768 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:58.975339+0000 mon.vm03 (mon.0) 769 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:58.975339+0000 mon.vm03 (mon.0) 769 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:59.028475+0000 mon.vm03 (mon.0) 770 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:59.028475+0000 mon.vm03 (mon.0) 770 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:59.034220+0000 mon.vm03 (mon.0) 771 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:59.034220+0000 mon.vm03 (mon.0) 771 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:59.035196+0000 mon.vm03 (mon.0) 772 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:59.035196+0000 mon.vm03 (mon.0) 772 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:59.035772+0000 mon.vm03 (mon.0) 773 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:59.035772+0000 mon.vm03 (mon.0) 773 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:59.040056+0000 mon.vm03 (mon.0) 774 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:59.040056+0000 mon.vm03 (mon.0) 774 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:59.041578+0000 mon.vm03 (mon.0) 775 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:59.041578+0000 mon.vm03 (mon.0) 775 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:59.048131+0000 mon.vm03 (mon.0) 776 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:20:59.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:20:59 vm08 bash[20828]: audit 2026-03-06T21:20:59.048131+0000 mon.vm03 (mon.0) 776 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' 2026-03-06T22:21:01.688 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:01 vm03 bash[17055]: cluster 2026-03-06T21:21:00.094217+0000 mgr.vm03.uwuzgl (mgr.14199) 212 : cluster [DBG] pgmap v137: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:01.688 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:01 vm03 bash[17055]: cluster 2026-03-06T21:21:00.094217+0000 mgr.vm03.uwuzgl (mgr.14199) 212 : cluster [DBG] pgmap v137: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:01.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:01 vm08 bash[20828]: cluster 2026-03-06T21:21:00.094217+0000 mgr.vm03.uwuzgl (mgr.14199) 212 : cluster [DBG] pgmap v137: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:01.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:01 vm08 bash[20828]: cluster 2026-03-06T21:21:00.094217+0000 mgr.vm03.uwuzgl (mgr.14199) 212 : cluster [DBG] pgmap v137: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:03.392 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:21:03.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:03 vm08 bash[20828]: cluster 2026-03-06T21:21:02.094515+0000 mgr.vm03.uwuzgl (mgr.14199) 213 : cluster [DBG] pgmap v138: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:03.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:03 vm08 bash[20828]: cluster 2026-03-06T21:21:02.094515+0000 mgr.vm03.uwuzgl (mgr.14199) 213 : cluster [DBG] pgmap v138: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:03.772 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-06T22:21:03.772 INFO:teuthology.orchestra.run.vm03.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-06T21:16:05.766418Z", "last_refresh": "2026-03-06T21:20:58.912498Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-06T21:17:13.981567Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-06T21:16:03.845298Z", "last_refresh": "2026-03-06T21:20:58.912305Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:14.763843Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-06T21:16:03.393952Z", "last_refresh": "2026-03-06T21:20:58.912276Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-06T21:16:04.987741Z", "last_refresh": "2026-03-06T21:20:58.912147Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-06T21:20:47.131435Z service:ingress.nfs.happy [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "nfs.happy", "service_name": "ingress.nfs.happy", "service_type": "ingress", "spec": {"backend_service": "nfs.happy", "enable_haproxy_protocol": true, "first_virtual_router_id": 50, "frontend_port": 2049, "monitor_port": 9049, "virtual_ip": "12.12.1.103"}, "status": {"created": "2026-03-06T21:20:21.847175Z", "last_refresh": "2026-03-06T21:20:58.912343Z", "ports": [2049, 9049], "running": 4, "size": 4, "virtual_ip": "12.12.1.103"}}, {"events": ["2026-03-06T21:20:20.634732Z service:mds.fs1 [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "fs1", "service_name": "mds.fs1", "service_type": "mds", "status": {"created": "2026-03-06T21:20:14.064886Z", "last_refresh": "2026-03-06T21:20:58.912385Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:16.189007Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T21:16:02.939076Z", "last_refresh": "2026-03-06T21:20:58.912690Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:17.196289Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm03:192.168.123.103=vm03", "vm08:192.168.123.108=vm08"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T21:17:07.022646Z", "last_refresh": "2026-03-06T21:20:58.912426Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T21:20:59.048383Z service:nfs.happy [INFO] \"service was created\""], "placement": {"count": 1}, "service_id": "happy", "service_name": "nfs.happy", "service_type": "nfs", "spec": {"enable_haproxy_protocol": true, "port": 12049}, "status": {"created": "2026-03-06T21:20:21.842604Z", "last_refresh": "2026-03-06T21:20:59.021683Z", "ports": [12049], "running": 1, "size": 1}}, {"events": ["2026-03-06T21:17:15.412676Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-06T21:16:05.375475Z", "last_refresh": "2026-03-06T21:20:58.912236Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-06T21:17:35.462286Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-06T21:17:35.457628Z", "last_refresh": "2026-03-06T21:20:58.912085Z", "running": 8, "size": 8}}, {"events": ["2026-03-06T21:17:17.200217Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-06T21:16:04.593189Z", "last_refresh": "2026-03-06T21:20:58.912616Z", "ports": [9095], "running": 1, "size": 1}}] 2026-03-06T22:21:03.784 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:03 vm03 bash[17055]: cluster 2026-03-06T21:21:02.094515+0000 mgr.vm03.uwuzgl (mgr.14199) 213 : cluster [DBG] pgmap v138: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:03.784 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:03 vm03 bash[17055]: cluster 2026-03-06T21:21:02.094515+0000 mgr.vm03.uwuzgl (mgr.14199) 213 : cluster [DBG] pgmap v138: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:03.841 INFO:tasks.cephadm:ingress.nfs.happy has 4/4 2026-03-06T22:21:03.841 INFO:teuthology.run_tasks:Running task vip.exec... 2026-03-06T22:21:03.844 INFO:tasks.vip:Running commands on role host.a host ubuntu@vm03.local 2026-03-06T22:21:03.844 DEBUG:teuthology.orchestra.run.vm03:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'mkdir /mnt/happy' 2026-03-06T22:21:03.851 INFO:teuthology.orchestra.run.vm03.stderr:+ mkdir /mnt/happy 2026-03-06T22:21:03.852 DEBUG:teuthology.orchestra.run.vm03:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'sleep 1' 2026-03-06T22:21:03.902 INFO:teuthology.orchestra.run.vm03.stderr:+ sleep 1 2026-03-06T22:21:04.904 DEBUG:teuthology.orchestra.run.vm03:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'mount -t nfs 12.12.1.103:/d1 /mnt/happy' 2026-03-06T22:21:04.954 INFO:teuthology.orchestra.run.vm03.stderr:+ mount -t nfs 12.12.1.103:/d1 /mnt/happy 2026-03-06T22:21:05.154 DEBUG:teuthology.orchestra.run.vm03:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'echo test > /mnt/happy/testfile' 2026-03-06T22:21:05.201 INFO:teuthology.orchestra.run.vm03.stderr:+ echo test 2026-03-06T22:21:05.212 DEBUG:teuthology.orchestra.run.vm03:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c sync 2026-03-06T22:21:05.260 INFO:teuthology.orchestra.run.vm03.stderr:+ sync 2026-03-06T22:21:05.266 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-06T22:21:05.268 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm03.local 2026-03-06T22:21:05.268 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- bash -c 'stat -c '"'"'%u %g'"'"' /var/log/ceph | grep '"'"'167 167'"'"'' 2026-03-06T22:21:05.688 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:05 vm03 bash[17055]: audit 2026-03-06T21:21:03.766282+0000 mgr.vm03.uwuzgl (mgr.14199) 214 : audit [DBG] from='client.14538 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:21:05.688 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:05 vm03 bash[17055]: audit 2026-03-06T21:21:03.766282+0000 mgr.vm03.uwuzgl (mgr.14199) 214 : audit [DBG] from='client.14538 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:21:05.688 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:05 vm03 bash[17055]: cluster 2026-03-06T21:21:04.094942+0000 mgr.vm03.uwuzgl (mgr.14199) 215 : cluster [DBG] pgmap v139: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 0 op/s 2026-03-06T22:21:05.688 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:05 vm03 bash[17055]: cluster 2026-03-06T21:21:04.094942+0000 mgr.vm03.uwuzgl (mgr.14199) 215 : cluster [DBG] pgmap v139: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 0 op/s 2026-03-06T22:21:05.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:05 vm08 bash[20828]: audit 2026-03-06T21:21:03.766282+0000 mgr.vm03.uwuzgl (mgr.14199) 214 : audit [DBG] from='client.14538 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:21:05.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:05 vm08 bash[20828]: audit 2026-03-06T21:21:03.766282+0000 mgr.vm03.uwuzgl (mgr.14199) 214 : audit [DBG] from='client.14538 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T22:21:05.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:05 vm08 bash[20828]: cluster 2026-03-06T21:21:04.094942+0000 mgr.vm03.uwuzgl (mgr.14199) 215 : cluster [DBG] pgmap v139: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 0 op/s 2026-03-06T22:21:05.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:05 vm08 bash[20828]: cluster 2026-03-06T21:21:04.094942+0000 mgr.vm03.uwuzgl (mgr.14199) 215 : cluster [DBG] pgmap v139: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 0 op/s 2026-03-06T22:21:06.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:06 vm08 bash[20828]: audit 2026-03-06T21:21:06.039641+0000 mon.vm03 (mon.0) 777 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.13", "id": [0, 4]}]: dispatch 2026-03-06T22:21:06.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:06 vm08 bash[20828]: audit 2026-03-06T21:21:06.039641+0000 mon.vm03 (mon.0) 777 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.13", "id": [0, 4]}]: dispatch 2026-03-06T22:21:06.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:06 vm08 bash[20828]: audit 2026-03-06T21:21:06.039843+0000 mon.vm03 (mon.0) 778 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.7", "id": [1, 2]}]: dispatch 2026-03-06T22:21:06.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:06 vm08 bash[20828]: audit 2026-03-06T21:21:06.039843+0000 mon.vm03 (mon.0) 778 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.7", "id": [1, 2]}]: dispatch 2026-03-06T22:21:06.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:06 vm08 bash[20828]: audit 2026-03-06T21:21:06.039925+0000 mon.vm03 (mon.0) 779 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.9", "id": [3, 7]}]: dispatch 2026-03-06T22:21:06.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:06 vm08 bash[20828]: audit 2026-03-06T21:21:06.039925+0000 mon.vm03 (mon.0) 779 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.9", "id": [3, 7]}]: dispatch 2026-03-06T22:21:06.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:06 vm08 bash[20828]: audit 2026-03-06T21:21:06.039979+0000 mon.vm03 (mon.0) 780 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.14", "id": [4, 5]}]: dispatch 2026-03-06T22:21:06.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:06 vm08 bash[20828]: audit 2026-03-06T21:21:06.039979+0000 mon.vm03 (mon.0) 780 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.14", "id": [4, 5]}]: dispatch 2026-03-06T22:21:06.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:06 vm08 bash[20828]: audit 2026-03-06T21:21:06.040032+0000 mon.vm03 (mon.0) 781 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.16", "id": [1, 6]}]: dispatch 2026-03-06T22:21:06.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:06 vm08 bash[20828]: audit 2026-03-06T21:21:06.040032+0000 mon.vm03 (mon.0) 781 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.16", "id": [1, 6]}]: dispatch 2026-03-06T22:21:06.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:06 vm08 bash[20828]: audit 2026-03-06T21:21:06.040086+0000 mon.vm03 (mon.0) 782 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.5", "id": [4, 2]}]: dispatch 2026-03-06T22:21:06.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:06 vm08 bash[20828]: audit 2026-03-06T21:21:06.040086+0000 mon.vm03 (mon.0) 782 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.5", "id": [4, 2]}]: dispatch 2026-03-06T22:21:06.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:06 vm08 bash[20828]: audit 2026-03-06T21:21:06.040138+0000 mon.vm03 (mon.0) 783 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.12", "id": [1, 2]}]: dispatch 2026-03-06T22:21:06.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:06 vm08 bash[20828]: audit 2026-03-06T21:21:06.040138+0000 mon.vm03 (mon.0) 783 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.12", "id": [1, 2]}]: dispatch 2026-03-06T22:21:06.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:06 vm08 bash[20828]: audit 2026-03-06T21:21:06.059687+0000 mon.vm03 (mon.0) 784 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:21:06.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:06 vm08 bash[20828]: audit 2026-03-06T21:21:06.059687+0000 mon.vm03 (mon.0) 784 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:21:06.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:06 vm03 bash[17055]: audit 2026-03-06T21:21:06.039641+0000 mon.vm03 (mon.0) 777 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.13", "id": [0, 4]}]: dispatch 2026-03-06T22:21:06.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:06 vm03 bash[17055]: audit 2026-03-06T21:21:06.039641+0000 mon.vm03 (mon.0) 777 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.13", "id": [0, 4]}]: dispatch 2026-03-06T22:21:06.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:06 vm03 bash[17055]: audit 2026-03-06T21:21:06.039843+0000 mon.vm03 (mon.0) 778 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.7", "id": [1, 2]}]: dispatch 2026-03-06T22:21:06.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:06 vm03 bash[17055]: audit 2026-03-06T21:21:06.039843+0000 mon.vm03 (mon.0) 778 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.7", "id": [1, 2]}]: dispatch 2026-03-06T22:21:06.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:06 vm03 bash[17055]: audit 2026-03-06T21:21:06.039925+0000 mon.vm03 (mon.0) 779 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.9", "id": [3, 7]}]: dispatch 2026-03-06T22:21:06.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:06 vm03 bash[17055]: audit 2026-03-06T21:21:06.039925+0000 mon.vm03 (mon.0) 779 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.9", "id": [3, 7]}]: dispatch 2026-03-06T22:21:06.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:06 vm03 bash[17055]: audit 2026-03-06T21:21:06.039979+0000 mon.vm03 (mon.0) 780 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.14", "id": [4, 5]}]: dispatch 2026-03-06T22:21:06.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:06 vm03 bash[17055]: audit 2026-03-06T21:21:06.039979+0000 mon.vm03 (mon.0) 780 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.14", "id": [4, 5]}]: dispatch 2026-03-06T22:21:06.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:06 vm03 bash[17055]: audit 2026-03-06T21:21:06.040032+0000 mon.vm03 (mon.0) 781 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.16", "id": [1, 6]}]: dispatch 2026-03-06T22:21:06.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:06 vm03 bash[17055]: audit 2026-03-06T21:21:06.040032+0000 mon.vm03 (mon.0) 781 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.16", "id": [1, 6]}]: dispatch 2026-03-06T22:21:06.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:06 vm03 bash[17055]: audit 2026-03-06T21:21:06.040086+0000 mon.vm03 (mon.0) 782 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.5", "id": [4, 2]}]: dispatch 2026-03-06T22:21:06.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:06 vm03 bash[17055]: audit 2026-03-06T21:21:06.040086+0000 mon.vm03 (mon.0) 782 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.5", "id": [4, 2]}]: dispatch 2026-03-06T22:21:06.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:06 vm03 bash[17055]: audit 2026-03-06T21:21:06.040138+0000 mon.vm03 (mon.0) 783 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.12", "id": [1, 2]}]: dispatch 2026-03-06T22:21:06.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:06 vm03 bash[17055]: audit 2026-03-06T21:21:06.040138+0000 mon.vm03 (mon.0) 783 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.12", "id": [1, 2]}]: dispatch 2026-03-06T22:21:06.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:06 vm03 bash[17055]: audit 2026-03-06T21:21:06.059687+0000 mon.vm03 (mon.0) 784 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:21:06.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:06 vm03 bash[17055]: audit 2026-03-06T21:21:06.059687+0000 mon.vm03 (mon.0) 784 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:21:07.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: cluster 2026-03-06T21:21:06.095226+0000 mgr.vm03.uwuzgl (mgr.14199) 216 : cluster [DBG] pgmap v140: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:07.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: cluster 2026-03-06T21:21:06.095226+0000 mgr.vm03.uwuzgl (mgr.14199) 216 : cluster [DBG] pgmap v140: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:07.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: audit 2026-03-06T21:21:06.451705+0000 mon.vm03 (mon.0) 785 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.13", "id": [0, 4]}]': finished 2026-03-06T22:21:07.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: audit 2026-03-06T21:21:06.451705+0000 mon.vm03 (mon.0) 785 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.13", "id": [0, 4]}]': finished 2026-03-06T22:21:07.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: audit 2026-03-06T21:21:06.451874+0000 mon.vm03 (mon.0) 786 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.7", "id": [1, 2]}]': finished 2026-03-06T22:21:07.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: audit 2026-03-06T21:21:06.451874+0000 mon.vm03 (mon.0) 786 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.7", "id": [1, 2]}]': finished 2026-03-06T22:21:07.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: audit 2026-03-06T21:21:06.451916+0000 mon.vm03 (mon.0) 787 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.9", "id": [3, 7]}]': finished 2026-03-06T22:21:07.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: audit 2026-03-06T21:21:06.451916+0000 mon.vm03 (mon.0) 787 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.9", "id": [3, 7]}]': finished 2026-03-06T22:21:07.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: audit 2026-03-06T21:21:06.451957+0000 mon.vm03 (mon.0) 788 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.14", "id": [4, 5]}]': finished 2026-03-06T22:21:07.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: audit 2026-03-06T21:21:06.451957+0000 mon.vm03 (mon.0) 788 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.14", "id": [4, 5]}]': finished 2026-03-06T22:21:07.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: audit 2026-03-06T21:21:06.451986+0000 mon.vm03 (mon.0) 789 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.16", "id": [1, 6]}]': finished 2026-03-06T22:21:07.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: audit 2026-03-06T21:21:06.451986+0000 mon.vm03 (mon.0) 789 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.16", "id": [1, 6]}]': finished 2026-03-06T22:21:07.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: audit 2026-03-06T21:21:06.452013+0000 mon.vm03 (mon.0) 790 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.5", "id": [4, 2]}]': finished 2026-03-06T22:21:07.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: audit 2026-03-06T21:21:06.452013+0000 mon.vm03 (mon.0) 790 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.5", "id": [4, 2]}]': finished 2026-03-06T22:21:07.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: audit 2026-03-06T21:21:06.452040+0000 mon.vm03 (mon.0) 791 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.12", "id": [1, 2]}]': finished 2026-03-06T22:21:07.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: audit 2026-03-06T21:21:06.452040+0000 mon.vm03 (mon.0) 791 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.12", "id": [1, 2]}]': finished 2026-03-06T22:21:07.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: cluster 2026-03-06T21:21:06.453493+0000 mon.vm03 (mon.0) 792 : cluster [DBG] osdmap e33: 8 total, 8 up, 8 in 2026-03-06T22:21:07.751 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:07 vm08 bash[20828]: cluster 2026-03-06T21:21:06.453493+0000 mon.vm03 (mon.0) 792 : cluster [DBG] osdmap e33: 8 total, 8 up, 8 in 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: cluster 2026-03-06T21:21:06.095226+0000 mgr.vm03.uwuzgl (mgr.14199) 216 : cluster [DBG] pgmap v140: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: cluster 2026-03-06T21:21:06.095226+0000 mgr.vm03.uwuzgl (mgr.14199) 216 : cluster [DBG] pgmap v140: 97 pgs: 97 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: audit 2026-03-06T21:21:06.451705+0000 mon.vm03 (mon.0) 785 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.13", "id": [0, 4]}]': finished 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: audit 2026-03-06T21:21:06.451705+0000 mon.vm03 (mon.0) 785 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.13", "id": [0, 4]}]': finished 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: audit 2026-03-06T21:21:06.451874+0000 mon.vm03 (mon.0) 786 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.7", "id": [1, 2]}]': finished 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: audit 2026-03-06T21:21:06.451874+0000 mon.vm03 (mon.0) 786 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.7", "id": [1, 2]}]': finished 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: audit 2026-03-06T21:21:06.451916+0000 mon.vm03 (mon.0) 787 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.9", "id": [3, 7]}]': finished 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: audit 2026-03-06T21:21:06.451916+0000 mon.vm03 (mon.0) 787 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.9", "id": [3, 7]}]': finished 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: audit 2026-03-06T21:21:06.451957+0000 mon.vm03 (mon.0) 788 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.14", "id": [4, 5]}]': finished 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: audit 2026-03-06T21:21:06.451957+0000 mon.vm03 (mon.0) 788 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.14", "id": [4, 5]}]': finished 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: audit 2026-03-06T21:21:06.451986+0000 mon.vm03 (mon.0) 789 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.16", "id": [1, 6]}]': finished 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: audit 2026-03-06T21:21:06.451986+0000 mon.vm03 (mon.0) 789 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.16", "id": [1, 6]}]': finished 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: audit 2026-03-06T21:21:06.452013+0000 mon.vm03 (mon.0) 790 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.5", "id": [4, 2]}]': finished 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: audit 2026-03-06T21:21:06.452013+0000 mon.vm03 (mon.0) 790 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.5", "id": [4, 2]}]': finished 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: audit 2026-03-06T21:21:06.452040+0000 mon.vm03 (mon.0) 791 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.12", "id": [1, 2]}]': finished 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: audit 2026-03-06T21:21:06.452040+0000 mon.vm03 (mon.0) 791 : audit [INF] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.12", "id": [1, 2]}]': finished 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: cluster 2026-03-06T21:21:06.453493+0000 mon.vm03 (mon.0) 792 : cluster [DBG] osdmap e33: 8 total, 8 up, 8 in 2026-03-06T22:21:07.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:07 vm03 bash[17055]: cluster 2026-03-06T21:21:06.453493+0000 mon.vm03 (mon.0) 792 : cluster [DBG] osdmap e33: 8 total, 8 up, 8 in 2026-03-06T22:21:08.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:08 vm08 bash[20828]: cluster 2026-03-06T21:21:07.459518+0000 mon.vm03 (mon.0) 793 : cluster [DBG] osdmap e34: 8 total, 8 up, 8 in 2026-03-06T22:21:08.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:08 vm08 bash[20828]: cluster 2026-03-06T21:21:07.459518+0000 mon.vm03 (mon.0) 793 : cluster [DBG] osdmap e34: 8 total, 8 up, 8 in 2026-03-06T22:21:08.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:08 vm03 bash[17055]: cluster 2026-03-06T21:21:07.459518+0000 mon.vm03 (mon.0) 793 : cluster [DBG] osdmap e34: 8 total, 8 up, 8 in 2026-03-06T22:21:08.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:08 vm03 bash[17055]: cluster 2026-03-06T21:21:07.459518+0000 mon.vm03 (mon.0) 793 : cluster [DBG] osdmap e34: 8 total, 8 up, 8 in 2026-03-06T22:21:09.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:09 vm08 bash[20828]: cluster 2026-03-06T21:21:08.095628+0000 mgr.vm03.uwuzgl (mgr.14199) 217 : cluster [DBG] pgmap v143: 97 pgs: 1 activating, 96 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 127 B/s rd, 127 B/s wr, 0 op/s 2026-03-06T22:21:09.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:09 vm08 bash[20828]: cluster 2026-03-06T21:21:08.095628+0000 mgr.vm03.uwuzgl (mgr.14199) 217 : cluster [DBG] pgmap v143: 97 pgs: 1 activating, 96 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 127 B/s rd, 127 B/s wr, 0 op/s 2026-03-06T22:21:09.937 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:09 vm03 bash[17055]: cluster 2026-03-06T21:21:08.095628+0000 mgr.vm03.uwuzgl (mgr.14199) 217 : cluster [DBG] pgmap v143: 97 pgs: 1 activating, 96 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 127 B/s rd, 127 B/s wr, 0 op/s 2026-03-06T22:21:09.938 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:09 vm03 bash[17055]: cluster 2026-03-06T21:21:08.095628+0000 mgr.vm03.uwuzgl (mgr.14199) 217 : cluster [DBG] pgmap v143: 97 pgs: 1 activating, 96 active+clean; 453 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 127 B/s rd, 127 B/s wr, 0 op/s 2026-03-06T22:21:10.142 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:21:10.242 INFO:teuthology.orchestra.run.vm03.stdout:167 167 2026-03-06T22:21:10.289 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- bash -c 'ceph orch status' 2026-03-06T22:21:11.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:11 vm08 bash[20828]: cluster 2026-03-06T21:21:10.095934+0000 mgr.vm03.uwuzgl (mgr.14199) 218 : cluster [DBG] pgmap v144: 97 pgs: 1 activating, 96 active+clean; 456 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 127 B/s rd, 1023 B/s wr, 0 op/s; 4 B/s, 1 keys/s, 0 objects/s recovering 2026-03-06T22:21:11.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:11 vm08 bash[20828]: cluster 2026-03-06T21:21:10.095934+0000 mgr.vm03.uwuzgl (mgr.14199) 218 : cluster [DBG] pgmap v144: 97 pgs: 1 activating, 96 active+clean; 456 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 127 B/s rd, 1023 B/s wr, 0 op/s; 4 B/s, 1 keys/s, 0 objects/s recovering 2026-03-06T22:21:11.937 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:11 vm03 bash[17055]: cluster 2026-03-06T21:21:10.095934+0000 mgr.vm03.uwuzgl (mgr.14199) 218 : cluster [DBG] pgmap v144: 97 pgs: 1 activating, 96 active+clean; 456 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 127 B/s rd, 1023 B/s wr, 0 op/s; 4 B/s, 1 keys/s, 0 objects/s recovering 2026-03-06T22:21:11.937 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:11 vm03 bash[17055]: cluster 2026-03-06T21:21:10.095934+0000 mgr.vm03.uwuzgl (mgr.14199) 218 : cluster [DBG] pgmap v144: 97 pgs: 1 activating, 96 active+clean; 456 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 127 B/s rd, 1023 B/s wr, 0 op/s; 4 B/s, 1 keys/s, 0 objects/s recovering 2026-03-06T22:21:13.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:13 vm08 bash[20828]: cluster 2026-03-06T21:21:12.096340+0000 mgr.vm03.uwuzgl (mgr.14199) 219 : cluster [DBG] pgmap v145: 97 pgs: 1 activating, 96 active+clean; 456 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1023 B/s wr, 0 op/s; 4 B/s, 1 keys/s, 0 objects/s recovering 2026-03-06T22:21:13.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:13 vm08 bash[20828]: cluster 2026-03-06T21:21:12.096340+0000 mgr.vm03.uwuzgl (mgr.14199) 219 : cluster [DBG] pgmap v145: 97 pgs: 1 activating, 96 active+clean; 456 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1023 B/s wr, 0 op/s; 4 B/s, 1 keys/s, 0 objects/s recovering 2026-03-06T22:21:13.937 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:13 vm03 bash[17055]: cluster 2026-03-06T21:21:12.096340+0000 mgr.vm03.uwuzgl (mgr.14199) 219 : cluster [DBG] pgmap v145: 97 pgs: 1 activating, 96 active+clean; 456 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1023 B/s wr, 0 op/s; 4 B/s, 1 keys/s, 0 objects/s recovering 2026-03-06T22:21:13.937 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:13 vm03 bash[17055]: cluster 2026-03-06T21:21:12.096340+0000 mgr.vm03.uwuzgl (mgr.14199) 219 : cluster [DBG] pgmap v145: 97 pgs: 1 activating, 96 active+clean; 456 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 1023 B/s wr, 0 op/s; 4 B/s, 1 keys/s, 0 objects/s recovering 2026-03-06T22:21:14.191 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:21:14.846 INFO:teuthology.orchestra.run.vm03.stdout:Backend: cephadm 2026-03-06T22:21:14.846 INFO:teuthology.orchestra.run.vm03.stdout:Available: Yes 2026-03-06T22:21:14.846 INFO:teuthology.orchestra.run.vm03.stdout:Paused: No 2026-03-06T22:21:14.913 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- bash -c 'ceph orch ps' 2026-03-06T22:21:16.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:15 vm08 bash[20828]: cluster 2026-03-06T21:21:14.096722+0000 mgr.vm03.uwuzgl (mgr.14199) 220 : cluster [DBG] pgmap v146: 97 pgs: 97 active+clean; 456 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 127 B/s rd, 1023 B/s wr, 1 op/s; 4 B/s, 1 keys/s, 0 objects/s recovering 2026-03-06T22:21:16.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:15 vm08 bash[20828]: cluster 2026-03-06T21:21:14.096722+0000 mgr.vm03.uwuzgl (mgr.14199) 220 : cluster [DBG] pgmap v146: 97 pgs: 97 active+clean; 456 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 127 B/s rd, 1023 B/s wr, 1 op/s; 4 B/s, 1 keys/s, 0 objects/s recovering 2026-03-06T22:21:16.187 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:15 vm03 bash[17055]: cluster 2026-03-06T21:21:14.096722+0000 mgr.vm03.uwuzgl (mgr.14199) 220 : cluster [DBG] pgmap v146: 97 pgs: 97 active+clean; 456 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 127 B/s rd, 1023 B/s wr, 1 op/s; 4 B/s, 1 keys/s, 0 objects/s recovering 2026-03-06T22:21:16.187 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:15 vm03 bash[17055]: cluster 2026-03-06T21:21:14.096722+0000 mgr.vm03.uwuzgl (mgr.14199) 220 : cluster [DBG] pgmap v146: 97 pgs: 97 active+clean; 456 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 127 B/s rd, 1023 B/s wr, 1 op/s; 4 B/s, 1 keys/s, 0 objects/s recovering 2026-03-06T22:21:17.187 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:16 vm03 bash[17055]: audit 2026-03-06T21:21:14.843742+0000 mgr.vm03.uwuzgl (mgr.14199) 221 : audit [DBG] from='client.14542 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:17.187 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:16 vm03 bash[17055]: audit 2026-03-06T21:21:14.843742+0000 mgr.vm03.uwuzgl (mgr.14199) 221 : audit [DBG] from='client.14542 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:17.187 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:16 vm03 bash[17055]: cluster 2026-03-06T21:21:16.096990+0000 mgr.vm03.uwuzgl (mgr.14199) 222 : cluster [DBG] pgmap v147: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 106 B/s rd, 1.0 KiB/s wr, 1 op/s; 3 B/s, 1 keys/s, 0 objects/s recovering 2026-03-06T22:21:17.187 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:16 vm03 bash[17055]: cluster 2026-03-06T21:21:16.096990+0000 mgr.vm03.uwuzgl (mgr.14199) 222 : cluster [DBG] pgmap v147: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 106 B/s rd, 1.0 KiB/s wr, 1 op/s; 3 B/s, 1 keys/s, 0 objects/s recovering 2026-03-06T22:21:17.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:16 vm08 bash[20828]: audit 2026-03-06T21:21:14.843742+0000 mgr.vm03.uwuzgl (mgr.14199) 221 : audit [DBG] from='client.14542 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:17.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:16 vm08 bash[20828]: audit 2026-03-06T21:21:14.843742+0000 mgr.vm03.uwuzgl (mgr.14199) 221 : audit [DBG] from='client.14542 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:17.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:16 vm08 bash[20828]: cluster 2026-03-06T21:21:16.096990+0000 mgr.vm03.uwuzgl (mgr.14199) 222 : cluster [DBG] pgmap v147: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 106 B/s rd, 1.0 KiB/s wr, 1 op/s; 3 B/s, 1 keys/s, 0 objects/s recovering 2026-03-06T22:21:17.251 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:16 vm08 bash[20828]: cluster 2026-03-06T21:21:16.096990+0000 mgr.vm03.uwuzgl (mgr.14199) 222 : cluster [DBG] pgmap v147: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 106 B/s rd, 1.0 KiB/s wr, 1 op/s; 3 B/s, 1 keys/s, 0 objects/s recovering 2026-03-06T22:21:19.437 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:19 vm03 bash[17055]: cluster 2026-03-06T21:21:18.097334+0000 mgr.vm03.uwuzgl (mgr.14199) 223 : cluster [DBG] pgmap v148: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 96 B/s rd, 962 B/s wr, 1 op/s; 3 B/s, 0 keys/s, 0 objects/s recovering 2026-03-06T22:21:19.437 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:19 vm03 bash[17055]: cluster 2026-03-06T21:21:18.097334+0000 mgr.vm03.uwuzgl (mgr.14199) 223 : cluster [DBG] pgmap v148: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 96 B/s rd, 962 B/s wr, 1 op/s; 3 B/s, 0 keys/s, 0 objects/s recovering 2026-03-06T22:21:19.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:19 vm08 bash[20828]: cluster 2026-03-06T21:21:18.097334+0000 mgr.vm03.uwuzgl (mgr.14199) 223 : cluster [DBG] pgmap v148: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 96 B/s rd, 962 B/s wr, 1 op/s; 3 B/s, 0 keys/s, 0 objects/s recovering 2026-03-06T22:21:19.502 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:19 vm08 bash[20828]: cluster 2026-03-06T21:21:18.097334+0000 mgr.vm03.uwuzgl (mgr.14199) 223 : cluster [DBG] pgmap v148: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 96 B/s rd, 962 B/s wr, 1 op/s; 3 B/s, 0 keys/s, 0 objects/s recovering 2026-03-06T22:21:19.707 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:alertmanager.vm03 vm03 *:9093,9094 running (3m) 21s ago 4m 14.7M - 0.25.0 c8568f914cd2 dfe36982fd82 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:ceph-exporter.vm03 vm03 running (4m) 21s ago 4m 9043k - 19.2.3-39-g340d3c24fc6 8bccc98d839a f09a38798cc8 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:ceph-exporter.vm08 vm08 running (4m) 21s ago 4m 6068k - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53153dc88f5a 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:crash.vm03 vm03 running (4m) 21s ago 4m 12.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 708f937ce517 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:crash.vm08 vm08 running (4m) 21s ago 4m 10.7M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 2bb962da32ac 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:grafana.vm03 vm03 *:3000 running (3m) 21s ago 4m 63.8M - 10.4.0 c8b91775d855 6bf553fcdf8b 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:haproxy.nfs.happy.vm03.agxegy vm03 *:2049,9049 running (43s) 21s ago 43s 3579k - 2.3.17-d1c9119 e85424b0d443 2192d4f51adb 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:haproxy.nfs.happy.vm08.kqaecw vm08 *:2049,9049 running (48s) 21s ago 48s 3556k - 2.3.17-d1c9119 e85424b0d443 eb80d2d14d43 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:keepalived.nfs.happy.vm03.vyaroa vm03 running (38s) 21s ago 38s 2480k - 2.2.4 4a3a1ff181d9 4f61d99d4ec7 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:keepalived.nfs.happy.vm08.rvikvg vm08 running (33s) 21s ago 33s 2496k - 2.2.4 4a3a1ff181d9 98740b42385e 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:mds.fs1.vm03.slersa vm03 running (60s) 21s ago 60s 13.4M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 8cdec02ffe7f 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:mds.fs1.vm08.gxmctn vm08 running (59s) 21s ago 59s 14.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a f34a972fc842 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:mgr.vm03.uwuzgl vm03 *:9283,8765,8443 running (5m) 21s ago 5m 539M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 446d797c0836 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:mgr.vm08.tdhano vm08 *:8443,9283,8765 running (4m) 21s ago 4m 473M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 18f9a0fe4daa 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:mon.vm03 vm03 running (5m) 21s ago 5m 53.5M 2048M 19.2.3-39-g340d3c24fc6 8bccc98d839a dedab1238dc2 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:mon.vm08 vm08 running (4m) 21s ago 4m 45.4M 2048M 19.2.3-39-g340d3c24fc6 8bccc98d839a e8de0b813933 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:nfs.happy.0.0.vm08.wdijel vm08 *:12049 running (53s) 21s ago 53s 50.0M - 5.9 8bccc98d839a 03bc3da4f5f3 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:node-exporter.vm03 vm03 *:9100 running (4m) 21s ago 4m 7656k - 1.7.0 72c9c2088986 d631da09166b 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:node-exporter.vm08 vm08 *:9100 running (4m) 21s ago 4m 7631k - 1.7.0 72c9c2088986 678b1f270c8e 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:osd.0 vm08 running (3m) 21s ago 3m 39.3M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 5f735e647563 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:osd.1 vm03 running (3m) 21s ago 3m 41.0M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 82414a348fdc 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:osd.2 vm08 running (3m) 21s ago 3m 39.6M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a a22a71708d83 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:osd.3 vm03 running (3m) 21s ago 3m 62.3M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a f968e8041eb8 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:osd.4 vm08 running (3m) 21s ago 3m 40.2M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a e1c933a67e93 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:osd.5 vm03 running (3m) 21s ago 3m 40.1M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a b50ce1f74566 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:osd.6 vm08 running (3m) 21s ago 3m 40.0M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 7a747b887e5b 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:osd.7 vm03 running (3m) 21s ago 3m 59.2M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 1e529bba7060 2026-03-06T22:21:20.105 INFO:teuthology.orchestra.run.vm03.stdout:prometheus.vm03 vm03 *:9095 running (26s) 21s ago 4m 32.6M - 2.51.0 1d3b7f56885b 07bf3746ffd9 2026-03-06T22:21:20.175 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- bash -c 'ceph orch ls' 2026-03-06T22:21:21.676 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:21 vm08 bash[20828]: audit 2026-03-06T21:21:20.096970+0000 mgr.vm03.uwuzgl (mgr.14199) 224 : audit [DBG] from='client.14546 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:21.676 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:21 vm08 bash[20828]: audit 2026-03-06T21:21:20.096970+0000 mgr.vm03.uwuzgl (mgr.14199) 224 : audit [DBG] from='client.14546 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:21.676 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:21 vm08 bash[20828]: cluster 2026-03-06T21:21:20.098340+0000 mgr.vm03.uwuzgl (mgr.14199) 225 : cluster [DBG] pgmap v149: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 767 B/s wr, 0 op/s; 2 B/s, 0 keys/s, 0 objects/s recovering 2026-03-06T22:21:21.676 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:21 vm08 bash[20828]: cluster 2026-03-06T21:21:20.098340+0000 mgr.vm03.uwuzgl (mgr.14199) 225 : cluster [DBG] pgmap v149: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 767 B/s wr, 0 op/s; 2 B/s, 0 keys/s, 0 objects/s recovering 2026-03-06T22:21:21.676 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:21 vm08 bash[20828]: audit 2026-03-06T21:21:21.060072+0000 mon.vm03 (mon.0) 794 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:21:21.676 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:21 vm08 bash[20828]: audit 2026-03-06T21:21:21.060072+0000 mon.vm03 (mon.0) 794 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:21:21.687 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:21 vm03 bash[17055]: audit 2026-03-06T21:21:20.096970+0000 mgr.vm03.uwuzgl (mgr.14199) 224 : audit [DBG] from='client.14546 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:21.687 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:21 vm03 bash[17055]: audit 2026-03-06T21:21:20.096970+0000 mgr.vm03.uwuzgl (mgr.14199) 224 : audit [DBG] from='client.14546 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:21.687 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:21 vm03 bash[17055]: cluster 2026-03-06T21:21:20.098340+0000 mgr.vm03.uwuzgl (mgr.14199) 225 : cluster [DBG] pgmap v149: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 767 B/s wr, 0 op/s; 2 B/s, 0 keys/s, 0 objects/s recovering 2026-03-06T22:21:21.687 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:21 vm03 bash[17055]: cluster 2026-03-06T21:21:20.098340+0000 mgr.vm03.uwuzgl (mgr.14199) 225 : cluster [DBG] pgmap v149: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 767 B/s wr, 0 op/s; 2 B/s, 0 keys/s, 0 objects/s recovering 2026-03-06T22:21:21.687 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:21 vm03 bash[17055]: audit 2026-03-06T21:21:21.060072+0000 mon.vm03 (mon.0) 794 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:21:21.687 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:21 vm03 bash[17055]: audit 2026-03-06T21:21:21.060072+0000 mon.vm03 (mon.0) 794 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:21:23.437 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:23 vm03 bash[17055]: cluster 2026-03-06T21:21:22.098698+0000 mgr.vm03.uwuzgl (mgr.14199) 226 : cluster [DBG] pgmap v150: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-06T22:21:23.437 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:23 vm03 bash[17055]: cluster 2026-03-06T21:21:22.098698+0000 mgr.vm03.uwuzgl (mgr.14199) 226 : cluster [DBG] pgmap v150: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-06T22:21:23.677 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:23 vm08 bash[20828]: cluster 2026-03-06T21:21:22.098698+0000 mgr.vm03.uwuzgl (mgr.14199) 226 : cluster [DBG] pgmap v150: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-06T22:21:23.677 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:23 vm08 bash[20828]: cluster 2026-03-06T21:21:22.098698+0000 mgr.vm03.uwuzgl (mgr.14199) 226 : cluster [DBG] pgmap v150: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-06T22:21:24.979 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:21:25.366 INFO:teuthology.orchestra.run.vm03.stdout:NAME PORTS RUNNING REFRESHED AGE PLACEMENT 2026-03-06T22:21:25.367 INFO:teuthology.orchestra.run.vm03.stdout:alertmanager ?:9093,9094 1/1 26s ago 5m count:1 2026-03-06T22:21:25.367 INFO:teuthology.orchestra.run.vm03.stdout:ceph-exporter 2/2 26s ago 5m * 2026-03-06T22:21:25.367 INFO:teuthology.orchestra.run.vm03.stdout:crash 2/2 26s ago 5m * 2026-03-06T22:21:25.367 INFO:teuthology.orchestra.run.vm03.stdout:grafana ?:3000 1/1 26s ago 5m count:1 2026-03-06T22:21:25.367 INFO:teuthology.orchestra.run.vm03.stdout:ingress.nfs.happy 12.12.1.103:2049,9049 4/4 26s ago 63s count:2 2026-03-06T22:21:25.367 INFO:teuthology.orchestra.run.vm03.stdout:mds.fs1 2/2 26s ago 71s count:2 2026-03-06T22:21:25.367 INFO:teuthology.orchestra.run.vm03.stdout:mgr 2/2 26s ago 5m count:2 2026-03-06T22:21:25.367 INFO:teuthology.orchestra.run.vm03.stdout:mon 2/2 26s ago 4m vm03:192.168.123.103=vm03;vm08:192.168.123.108=vm08;count:2 2026-03-06T22:21:25.367 INFO:teuthology.orchestra.run.vm03.stdout:nfs.happy ?:12049 1/1 26s ago 63s count:1 2026-03-06T22:21:25.367 INFO:teuthology.orchestra.run.vm03.stdout:node-exporter ?:9100 2/2 26s ago 5m * 2026-03-06T22:21:25.367 INFO:teuthology.orchestra.run.vm03.stdout:osd.all-available-devices 8 26s ago 3m * 2026-03-06T22:21:25.367 INFO:teuthology.orchestra.run.vm03.stdout:prometheus ?:9095 1/1 26s ago 5m count:1 2026-03-06T22:21:25.444 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- bash -c 'ceph orch host ls' 2026-03-06T22:21:25.679 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:25 vm08 bash[20828]: cluster 2026-03-06T21:21:24.099267+0000 mgr.vm03.uwuzgl (mgr.14199) 227 : cluster [DBG] pgmap v151: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-06T22:21:25.679 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:25 vm08 bash[20828]: cluster 2026-03-06T21:21:24.099267+0000 mgr.vm03.uwuzgl (mgr.14199) 227 : cluster [DBG] pgmap v151: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-06T22:21:25.686 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:25 vm03 bash[17055]: cluster 2026-03-06T21:21:24.099267+0000 mgr.vm03.uwuzgl (mgr.14199) 227 : cluster [DBG] pgmap v151: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-06T22:21:25.686 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:25 vm03 bash[17055]: cluster 2026-03-06T21:21:24.099267+0000 mgr.vm03.uwuzgl (mgr.14199) 227 : cluster [DBG] pgmap v151: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-06T22:21:26.686 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:26 vm03 bash[17055]: audit 2026-03-06T21:21:25.361077+0000 mgr.vm03.uwuzgl (mgr.14199) 228 : audit [DBG] from='client.14550 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:26.686 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:26 vm03 bash[17055]: audit 2026-03-06T21:21:25.361077+0000 mgr.vm03.uwuzgl (mgr.14199) 228 : audit [DBG] from='client.14550 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:26.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:26 vm08 bash[20828]: audit 2026-03-06T21:21:25.361077+0000 mgr.vm03.uwuzgl (mgr.14199) 228 : audit [DBG] from='client.14550 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:26.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:26 vm08 bash[20828]: audit 2026-03-06T21:21:25.361077+0000 mgr.vm03.uwuzgl (mgr.14199) 228 : audit [DBG] from='client.14550 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:27.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:27 vm08 bash[20828]: cluster 2026-03-06T21:21:26.099587+0000 mgr.vm03.uwuzgl (mgr.14199) 229 : cluster [DBG] pgmap v152: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-06T22:21:27.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:27 vm08 bash[20828]: cluster 2026-03-06T21:21:26.099587+0000 mgr.vm03.uwuzgl (mgr.14199) 229 : cluster [DBG] pgmap v152: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-06T22:21:27.784 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:27 vm03 bash[17055]: cluster 2026-03-06T21:21:26.099587+0000 mgr.vm03.uwuzgl (mgr.14199) 229 : cluster [DBG] pgmap v152: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-06T22:21:27.784 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:27 vm03 bash[17055]: cluster 2026-03-06T21:21:26.099587+0000 mgr.vm03.uwuzgl (mgr.14199) 229 : cluster [DBG] pgmap v152: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-06T22:21:29.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:29 vm08 bash[20828]: cluster 2026-03-06T21:21:28.099939+0000 mgr.vm03.uwuzgl (mgr.14199) 230 : cluster [DBG] pgmap v153: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:29.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:29 vm08 bash[20828]: cluster 2026-03-06T21:21:28.099939+0000 mgr.vm03.uwuzgl (mgr.14199) 230 : cluster [DBG] pgmap v153: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:29.936 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:29 vm03 bash[17055]: cluster 2026-03-06T21:21:28.099939+0000 mgr.vm03.uwuzgl (mgr.14199) 230 : cluster [DBG] pgmap v153: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:29.936 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:29 vm03 bash[17055]: cluster 2026-03-06T21:21:28.099939+0000 mgr.vm03.uwuzgl (mgr.14199) 230 : cluster [DBG] pgmap v153: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:30.278 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:21:30.650 INFO:teuthology.orchestra.run.vm03.stdout:HOST ADDR LABELS STATUS 2026-03-06T22:21:30.650 INFO:teuthology.orchestra.run.vm03.stdout:vm03 192.168.123.103 2026-03-06T22:21:30.650 INFO:teuthology.orchestra.run.vm03.stdout:vm08 192.168.123.108 2026-03-06T22:21:30.650 INFO:teuthology.orchestra.run.vm03.stdout:2 hosts in cluster 2026-03-06T22:21:30.721 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- bash -c 'ceph orch device ls' 2026-03-06T22:21:31.686 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:31 vm03 bash[17055]: cluster 2026-03-06T21:21:30.100361+0000 mgr.vm03.uwuzgl (mgr.14199) 231 : cluster [DBG] pgmap v154: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:31.686 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:31 vm03 bash[17055]: cluster 2026-03-06T21:21:30.100361+0000 mgr.vm03.uwuzgl (mgr.14199) 231 : cluster [DBG] pgmap v154: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:31.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:31 vm08 bash[20828]: cluster 2026-03-06T21:21:30.100361+0000 mgr.vm03.uwuzgl (mgr.14199) 231 : cluster [DBG] pgmap v154: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:31.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:31 vm08 bash[20828]: cluster 2026-03-06T21:21:30.100361+0000 mgr.vm03.uwuzgl (mgr.14199) 231 : cluster [DBG] pgmap v154: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:32.686 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:32 vm03 bash[17055]: audit 2026-03-06T21:21:30.647513+0000 mgr.vm03.uwuzgl (mgr.14199) 232 : audit [DBG] from='client.14554 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:32.686 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:32 vm03 bash[17055]: audit 2026-03-06T21:21:30.647513+0000 mgr.vm03.uwuzgl (mgr.14199) 232 : audit [DBG] from='client.14554 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:32.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:32 vm08 bash[20828]: audit 2026-03-06T21:21:30.647513+0000 mgr.vm03.uwuzgl (mgr.14199) 232 : audit [DBG] from='client.14554 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:32.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:32 vm08 bash[20828]: audit 2026-03-06T21:21:30.647513+0000 mgr.vm03.uwuzgl (mgr.14199) 232 : audit [DBG] from='client.14554 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:33.686 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:33 vm03 bash[17055]: cluster 2026-03-06T21:21:32.100629+0000 mgr.vm03.uwuzgl (mgr.14199) 233 : cluster [DBG] pgmap v155: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:33.686 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:33 vm03 bash[17055]: cluster 2026-03-06T21:21:32.100629+0000 mgr.vm03.uwuzgl (mgr.14199) 233 : cluster [DBG] pgmap v155: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:33.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:33 vm08 bash[20828]: cluster 2026-03-06T21:21:32.100629+0000 mgr.vm03.uwuzgl (mgr.14199) 233 : cluster [DBG] pgmap v155: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:33.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:33 vm08 bash[20828]: cluster 2026-03-06T21:21:32.100629+0000 mgr.vm03.uwuzgl (mgr.14199) 233 : cluster [DBG] pgmap v155: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:35.537 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:21:35.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:35 vm08 bash[20828]: cluster 2026-03-06T21:21:34.101052+0000 mgr.vm03.uwuzgl (mgr.14199) 234 : cluster [DBG] pgmap v156: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 0 op/s 2026-03-06T22:21:35.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:35 vm08 bash[20828]: cluster 2026-03-06T21:21:34.101052+0000 mgr.vm03.uwuzgl (mgr.14199) 234 : cluster [DBG] pgmap v156: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 0 op/s 2026-03-06T22:21:35.936 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:35 vm03 bash[17055]: cluster 2026-03-06T21:21:34.101052+0000 mgr.vm03.uwuzgl (mgr.14199) 234 : cluster [DBG] pgmap v156: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 0 op/s 2026-03-06T22:21:35.936 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:35 vm03 bash[17055]: cluster 2026-03-06T21:21:34.101052+0000 mgr.vm03.uwuzgl (mgr.14199) 234 : cluster [DBG] pgmap v156: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 0 op/s 2026-03-06T22:21:35.941 INFO:teuthology.orchestra.run.vm03.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REFRESHED REJECT REASONS 2026-03-06T22:21:35.941 INFO:teuthology.orchestra.run.vm03.stdout:vm03 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 77s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-06T22:21:35.941 INFO:teuthology.orchestra.run.vm03.stdout:vm03 /dev/vdb hdd DWNBRSTVMM03001 20.0G No 77s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:21:35.941 INFO:teuthology.orchestra.run.vm03.stdout:vm03 /dev/vdc hdd DWNBRSTVMM03002 20.0G No 77s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:21:35.941 INFO:teuthology.orchestra.run.vm03.stdout:vm03 /dev/vdd hdd DWNBRSTVMM03003 20.0G No 77s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:21:35.941 INFO:teuthology.orchestra.run.vm03.stdout:vm03 /dev/vde hdd DWNBRSTVMM03004 20.0G No 77s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:21:35.941 INFO:teuthology.orchestra.run.vm03.stdout:vm08 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 77s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-06T22:21:35.941 INFO:teuthology.orchestra.run.vm03.stdout:vm08 /dev/vdb hdd DWNBRSTVMM08001 20.0G No 77s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:21:35.941 INFO:teuthology.orchestra.run.vm03.stdout:vm08 /dev/vdc hdd DWNBRSTVMM08002 20.0G No 77s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:21:35.941 INFO:teuthology.orchestra.run.vm03.stdout:vm08 /dev/vdd hdd DWNBRSTVMM08003 20.0G No 77s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:21:35.941 INFO:teuthology.orchestra.run.vm03.stdout:vm08 /dev/vde hdd DWNBRSTVMM08004 20.0G No 77s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-06T22:21:36.016 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 -- bash -c 'ceph orch ls | grep '"'"'^osd.all-available-devices '"'"'' 2026-03-06T22:21:36.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:36 vm08 bash[20828]: audit 2026-03-06T21:21:36.060164+0000 mon.vm03 (mon.0) 795 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:21:36.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:36 vm08 bash[20828]: audit 2026-03-06T21:21:36.060164+0000 mon.vm03 (mon.0) 795 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:21:36.936 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:36 vm03 bash[17055]: audit 2026-03-06T21:21:36.060164+0000 mon.vm03 (mon.0) 795 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:21:36.936 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:36 vm03 bash[17055]: audit 2026-03-06T21:21:36.060164+0000 mon.vm03 (mon.0) 795 : audit [DBG] from='mgr.14199 192.168.123.103:0/3318381354' entity='mgr.vm03.uwuzgl' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-06T22:21:37.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:37 vm08 bash[20828]: audit 2026-03-06T21:21:35.938949+0000 mgr.vm03.uwuzgl (mgr.14199) 235 : audit [DBG] from='client.14558 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:37.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:37 vm08 bash[20828]: audit 2026-03-06T21:21:35.938949+0000 mgr.vm03.uwuzgl (mgr.14199) 235 : audit [DBG] from='client.14558 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:37.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:37 vm08 bash[20828]: cluster 2026-03-06T21:21:36.101431+0000 mgr.vm03.uwuzgl (mgr.14199) 236 : cluster [DBG] pgmap v157: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:37.750 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:37 vm08 bash[20828]: cluster 2026-03-06T21:21:36.101431+0000 mgr.vm03.uwuzgl (mgr.14199) 236 : cluster [DBG] pgmap v157: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:37.936 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:37 vm03 bash[17055]: audit 2026-03-06T21:21:35.938949+0000 mgr.vm03.uwuzgl (mgr.14199) 235 : audit [DBG] from='client.14558 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:37.936 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:37 vm03 bash[17055]: audit 2026-03-06T21:21:35.938949+0000 mgr.vm03.uwuzgl (mgr.14199) 235 : audit [DBG] from='client.14558 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T22:21:37.936 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:37 vm03 bash[17055]: cluster 2026-03-06T21:21:36.101431+0000 mgr.vm03.uwuzgl (mgr.14199) 236 : cluster [DBG] pgmap v157: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:37.936 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:37 vm03 bash[17055]: cluster 2026-03-06T21:21:36.101431+0000 mgr.vm03.uwuzgl (mgr.14199) 236 : cluster [DBG] pgmap v157: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:39.936 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:39 vm03 bash[17055]: cluster 2026-03-06T21:21:38.101755+0000 mgr.vm03.uwuzgl (mgr.14199) 237 : cluster [DBG] pgmap v158: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:39.936 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:39 vm03 bash[17055]: cluster 2026-03-06T21:21:38.101755+0000 mgr.vm03.uwuzgl (mgr.14199) 237 : cluster [DBG] pgmap v158: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:40.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:39 vm08 bash[20828]: cluster 2026-03-06T21:21:38.101755+0000 mgr.vm03.uwuzgl (mgr.14199) 237 : cluster [DBG] pgmap v158: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:40.000 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:39 vm08 bash[20828]: cluster 2026-03-06T21:21:38.101755+0000 mgr.vm03.uwuzgl (mgr.14199) 237 : cluster [DBG] pgmap v158: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:40.872 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/mon.vm03/config 2026-03-06T22:21:41.257 INFO:teuthology.orchestra.run.vm03.stdout:osd.all-available-devices 8 42s ago 4m * 2026-03-06T22:21:41.318 DEBUG:teuthology.run_tasks:Unwinding manager vip 2026-03-06T22:21:41.321 INFO:tasks.vip:Removing 12.12.0.103 (and any VIPs) on vm03.local iface ens3... 2026-03-06T22:21:41.321 DEBUG:teuthology.orchestra.run.vm03:> sudo ip addr del 12.12.0.103/22 dev ens3 2026-03-06T22:21:41.331 DEBUG:teuthology.orchestra.run.vm03:> sudo ip addr del 12.12.1.103/22 dev ens3 2026-03-06T22:21:41.379 INFO:teuthology.orchestra.run.vm03.stderr:RTNETLINK answers: Cannot assign requested address 2026-03-06T22:21:41.380 DEBUG:teuthology.orchestra.run:got remote process result: 2 2026-03-06T22:21:41.380 INFO:tasks.vip:Removing 12.12.0.108 (and any VIPs) on vm08.local iface ens3... 2026-03-06T22:21:41.380 DEBUG:teuthology.orchestra.run.vm08:> sudo ip addr del 12.12.0.108/22 dev ens3 2026-03-06T22:21:41.388 DEBUG:teuthology.orchestra.run.vm08:> sudo ip addr del 12.12.1.103/22 dev ens3 2026-03-06T22:21:41.434 INFO:teuthology.orchestra.run.vm08.stderr:RTNETLINK answers: Cannot assign requested address 2026-03-06T22:21:41.434 DEBUG:teuthology.orchestra.run:got remote process result: 2 2026-03-06T22:21:41.434 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-06T22:21:41.436 INFO:tasks.cephadm:Teardown begin 2026-03-06T22:21:41.437 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T22:21:41.445 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T22:21:41.482 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-06T22:21:41.482 DEBUG:teuthology.orchestra.run.vm03:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-06T22:21:41.492 DEBUG:teuthology.orchestra.run.vm08:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-06T22:21:41.531 INFO:tasks.cephadm:Stopping all daemons... 2026-03-06T22:21:41.531 INFO:tasks.cephadm.mon.vm03:Stopping mon.vm03... 2026-03-06T22:21:41.531 DEBUG:teuthology.orchestra.run.vm03:> sudo systemctl stop ceph-894e000c-19a1-11f1-8dbe-23b24380a082@mon.vm03 2026-03-06T22:21:41.538 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:41 vm03 bash[17055]: cluster 2026-03-06T21:21:40.102168+0000 mgr.vm03.uwuzgl (mgr.14199) 238 : cluster [DBG] pgmap v159: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:41.538 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:41 vm03 bash[17055]: cluster 2026-03-06T21:21:40.102168+0000 mgr.vm03.uwuzgl (mgr.14199) 238 : cluster [DBG] pgmap v159: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:41.582 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:41 vm03 systemd[1]: Stopping Ceph mon.vm03 for 894e000c-19a1-11f1-8dbe-23b24380a082... 2026-03-06T22:21:41.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:41 vm08 bash[20828]: cluster 2026-03-06T21:21:40.102168+0000 mgr.vm03.uwuzgl (mgr.14199) 238 : cluster [DBG] pgmap v159: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:41.693 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:41 vm08 bash[20828]: cluster 2026-03-06T21:21:40.102168+0000 mgr.vm03.uwuzgl (mgr.14199) 238 : cluster [DBG] pgmap v159: 97 pgs: 97 active+clean; 458 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 0 op/s 2026-03-06T22:21:41.885 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:41 vm03 bash[17055]: debug 2026-03-06T21:21:41.577+0000 7ff81b84c640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.vm03 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-06T22:21:41.885 INFO:journalctl@ceph.mon.vm03.vm03.stdout:Mar 06 22:21:41 vm03 bash[17055]: debug 2026-03-06T21:21:41.577+0000 7ff81b84c640 -1 mon.vm03@0(leader) e2 *** Got Signal Terminated *** 2026-03-06T22:21:41.946 DEBUG:teuthology.orchestra.run.vm03:> sudo pkill -f 'journalctl -f -n 0 -u ceph-894e000c-19a1-11f1-8dbe-23b24380a082@mon.vm03.service' 2026-03-06T22:21:42.014 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T22:21:42.014 INFO:tasks.cephadm.mon.vm03:Stopped mon.vm03 2026-03-06T22:21:42.014 INFO:tasks.cephadm.mon.vm08:Stopping mon.vm08... 2026-03-06T22:21:42.014 DEBUG:teuthology.orchestra.run.vm08:> sudo systemctl stop ceph-894e000c-19a1-11f1-8dbe-23b24380a082@mon.vm08 2026-03-06T22:21:42.282 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:42 vm08 systemd[1]: Stopping Ceph mon.vm08 for 894e000c-19a1-11f1-8dbe-23b24380a082... 2026-03-06T22:21:42.282 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:42 vm08 bash[20828]: debug 2026-03-06T21:21:42.071+0000 7fd0a1dc9640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.vm08 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-06T22:21:42.282 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 06 22:21:42 vm08 bash[20828]: debug 2026-03-06T21:21:42.071+0000 7fd0a1dc9640 -1 mon.vm08@1(peon) e2 *** Got Signal Terminated *** 2026-03-06T22:21:42.361 DEBUG:teuthology.orchestra.run.vm08:> sudo pkill -f 'journalctl -f -n 0 -u ceph-894e000c-19a1-11f1-8dbe-23b24380a082@mon.vm08.service' 2026-03-06T22:21:42.390 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T22:21:42.390 INFO:tasks.cephadm.mon.vm08:Stopped mon.vm08 2026-03-06T22:21:42.390 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 --force --keep-logs 2026-03-06T22:21:42.662 INFO:teuthology.orchestra.run.vm03.stdout:Deleting cluster with fsid: 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:22:17.020 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 --force --keep-logs 2026-03-06T22:22:17.275 INFO:teuthology.orchestra.run.vm08.stdout:Deleting cluster with fsid: 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:23:07.124 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T22:23:07.131 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T22:23:07.138 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-06T22:23:07.138 DEBUG:teuthology.misc:Transferring archived files from vm03:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/crash to /archive/irq0-2026-03-06_20:21:59-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/382/remote/vm03/crash 2026-03-06T22:23:07.138 DEBUG:teuthology.orchestra.run.vm03:> sudo tar c -f - -C /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/crash -- . 2026-03-06T22:23:07.180 INFO:teuthology.orchestra.run.vm03.stderr:tar: /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/crash: Cannot open: No such file or directory 2026-03-06T22:23:07.180 INFO:teuthology.orchestra.run.vm03.stderr:tar: Error is not recoverable: exiting now 2026-03-06T22:23:07.181 DEBUG:teuthology.misc:Transferring archived files from vm08:/var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/crash to /archive/irq0-2026-03-06_20:21:59-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/382/remote/vm08/crash 2026-03-06T22:23:07.181 DEBUG:teuthology.orchestra.run.vm08:> sudo tar c -f - -C /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/crash -- . 2026-03-06T22:23:07.187 INFO:teuthology.orchestra.run.vm08.stderr:tar: /var/lib/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/crash: Cannot open: No such file or directory 2026-03-06T22:23:07.187 INFO:teuthology.orchestra.run.vm08.stderr:tar: Error is not recoverable: exiting now 2026-03-06T22:23:07.188 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-06T22:23:07.188 DEBUG:teuthology.orchestra.run.vm03:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v CEPHADM_DAEMON_PLACE_FAIL | egrep -v CEPHADM_FAILED_DAEMON | head -n 1 2026-03-06T22:23:07.231 INFO:tasks.cephadm:Compressing logs... 2026-03-06T22:23:07.231 DEBUG:teuthology.orchestra.run.vm03:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T22:23:07.272 DEBUG:teuthology.orchestra.run.vm08:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T22:23:07.277 INFO:teuthology.orchestra.run.vm03.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-06T22:23:07.279 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-06T22:23:07.279 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.3.log 2026-03-06T22:23:07.279 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.log 2026-03-06T22:23:07.280 INFO:teuthology.orchestra.run.vm08.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-06T22:23:07.281 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-06T22:23:07.281 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mgr.vm08.tdhano.log 2026-03-06T22:23:07.281 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.log 2026-03-06T22:23:07.282 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mgr.vm08.tdhano.log: 89.5% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-06T22:23:07.283 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-client.ceph-exporter.vm08.log 2026-03-06T22:23:07.283 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.log: 86.3% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.log.gz 2026-03-06T22:23:07.284 INFO:teuthology.orchestra.run.vm08.stderr: 91.6%gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.6.log 2026-03-06T22:23:07.284 INFO:teuthology.orchestra.run.vm08.stderr: -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mgr.vm08.tdhano.log.gz 2026-03-06T22:23:07.284 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-client.ceph-exporter.vm08.log: 30.0% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-client.ceph-exporter.vm08.log.gz 2026-03-06T22:23:07.284 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.2.log 2026-03-06T22:23:07.284 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.audit.log 2026-03-06T22:23:07.291 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.3.log: gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mgr.vm03.uwuzgl.log 2026-03-06T22:23:07.292 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.log: 86.3% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.log.gz 2026-03-06T22:23:07.293 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.6.log: /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.2.log: gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-volume.log 2026-03-06T22:23:07.293 INFO:teuthology.orchestra.run.vm03.stderr: 91.2% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-06T22:23:07.293 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.1.log 2026-03-06T22:23:07.294 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.audit.log: 90.9% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.audit.log.gz 2026-03-06T22:23:07.297 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mon.vm08.log 2026-03-06T22:23:07.299 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mgr.vm03.uwuzgl.log: gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-client.ceph-exporter.vm03.log 2026-03-06T22:23:07.311 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.1.log: gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.5.log 2026-03-06T22:23:07.312 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-client.ceph-exporter.vm03.log: 93.9% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-client.ceph-exporter.vm03.log.gz 2026-03-06T22:23:07.313 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-volume.log: gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mds.fs1.vm08.gxmctn.log 2026-03-06T22:23:07.317 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mon.vm08.log: gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.cephadm.log 2026-03-06T22:23:07.321 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mds.fs1.vm08.gxmctn.log: 85.2% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mds.fs1.vm08.gxmctn.log.gz 2026-03-06T22:23:07.321 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.4.log 2026-03-06T22:23:07.321 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.cephadm.log: 81.8% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.cephadm.log.gz 2026-03-06T22:23:07.328 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.7.log 2026-03-06T22:23:07.336 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.5.log: gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.audit.log 2026-03-06T22:23:07.339 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.7.log: gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-volume.log 2026-03-06T22:23:07.341 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.0.log 2026-03-06T22:23:07.344 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.audit.log: gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mds.fs1.vm03.slersa.log 2026-03-06T22:23:07.345 INFO:teuthology.orchestra.run.vm03.stderr: 90.8% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.audit.log.gz 2026-03-06T22:23:07.352 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.cephadm.log 2026-03-06T22:23:07.352 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mds.fs1.vm03.slersa.log: /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-volume.log: 68.1% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mds.fs1.vm03.slersa.log.gz 2026-03-06T22:23:07.352 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mon.vm03.log 2026-03-06T22:23:07.356 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.cephadm.log: 82.8% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph.cephadm.log.gz 2026-03-06T22:23:07.361 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.4.log: /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.0.log: 93.3% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-volume.log.gz 2026-03-06T22:23:07.373 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mon.vm03.log: 93.7% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.3.log.gz 2026-03-06T22:23:07.405 INFO:teuthology.orchestra.run.vm08.stderr: 93.5% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.6.log.gz 2026-03-06T22:23:07.420 INFO:teuthology.orchestra.run.vm03.stderr: 93.2% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-volume.log.gz 2026-03-06T22:23:07.422 INFO:teuthology.orchestra.run.vm08.stderr: 92.6% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mon.vm08.log.gz 2026-03-06T22:23:07.444 INFO:teuthology.orchestra.run.vm08.stderr: 93.1% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.2.log.gz 2026-03-06T22:23:07.453 INFO:teuthology.orchestra.run.vm08.stderr: 93.5% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.4.log.gz 2026-03-06T22:23:07.457 INFO:teuthology.orchestra.run.vm03.stderr: 93.1% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.7.log.gz 2026-03-06T22:23:07.472 INFO:teuthology.orchestra.run.vm08.stderr: 93.6% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.0.log.gz 2026-03-06T22:23:07.473 INFO:teuthology.orchestra.run.vm08.stderr: 2026-03-06T22:23:07.473 INFO:teuthology.orchestra.run.vm08.stderr:real 0m0.198s 2026-03-06T22:23:07.473 INFO:teuthology.orchestra.run.vm08.stderr:user 0m0.341s 2026-03-06T22:23:07.473 INFO:teuthology.orchestra.run.vm08.stderr:sys 0m0.030s 2026-03-06T22:23:07.488 INFO:teuthology.orchestra.run.vm03.stderr: 90.1% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mgr.vm03.uwuzgl.log.gz 2026-03-06T22:23:07.491 INFO:teuthology.orchestra.run.vm03.stderr: 93.5% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.5.log.gz 2026-03-06T22:23:07.500 INFO:teuthology.orchestra.run.vm03.stderr: 94.1% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-osd.1.log.gz 2026-03-06T22:23:07.557 INFO:teuthology.orchestra.run.vm03.stderr: 91.4% -- replaced with /var/log/ceph/894e000c-19a1-11f1-8dbe-23b24380a082/ceph-mon.vm03.log.gz 2026-03-06T22:23:07.559 INFO:teuthology.orchestra.run.vm03.stderr: 2026-03-06T22:23:07.559 INFO:teuthology.orchestra.run.vm03.stderr:real 0m0.286s 2026-03-06T22:23:07.559 INFO:teuthology.orchestra.run.vm03.stderr:user 0m0.472s 2026-03-06T22:23:07.559 INFO:teuthology.orchestra.run.vm03.stderr:sys 0m0.034s 2026-03-06T22:23:07.559 INFO:tasks.cephadm:Archiving logs... 2026-03-06T22:23:07.559 DEBUG:teuthology.misc:Transferring archived files from vm03:/var/log/ceph to /archive/irq0-2026-03-06_20:21:59-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/382/remote/vm03/log 2026-03-06T22:23:07.559 DEBUG:teuthology.orchestra.run.vm03:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-06T22:23:07.643 DEBUG:teuthology.misc:Transferring archived files from vm08:/var/log/ceph to /archive/irq0-2026-03-06_20:21:59-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/382/remote/vm08/log 2026-03-06T22:23:07.643 DEBUG:teuthology.orchestra.run.vm08:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-06T22:23:07.675 INFO:tasks.cephadm:Removing cluster... 2026-03-06T22:23:07.675 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 --force 2026-03-06T22:23:07.926 INFO:teuthology.orchestra.run.vm03.stdout:Deleting cluster with fsid: 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:23:08.999 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 894e000c-19a1-11f1-8dbe-23b24380a082 --force 2026-03-06T22:23:09.262 INFO:teuthology.orchestra.run.vm08.stdout:Deleting cluster with fsid: 894e000c-19a1-11f1-8dbe-23b24380a082 2026-03-06T22:23:10.340 INFO:tasks.cephadm:Removing cephadm ... 2026-03-06T22:23:10.340 DEBUG:teuthology.orchestra.run.vm03:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-06T22:23:10.344 DEBUG:teuthology.orchestra.run.vm08:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-06T22:23:10.348 INFO:tasks.cephadm:Teardown complete 2026-03-06T22:23:10.348 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-06T22:23:10.350 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-06T22:23:10.350 DEBUG:teuthology.orchestra.run.vm03:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T22:23:10.388 DEBUG:teuthology.orchestra.run.vm08:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T22:23:11.101 INFO:teuthology.orchestra.run.vm03.stdout: remote refid st t when poll reach delay offset jitter 2026-03-06T22:23:11.101 INFO:teuthology.orchestra.run.vm03.stdout:============================================================================== 2026-03-06T22:23:11.101 INFO:teuthology.orchestra.run.vm03.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:23:11.101 INFO:teuthology.orchestra.run.vm03.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:23:11.101 INFO:teuthology.orchestra.run.vm03.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:23:11.101 INFO:teuthology.orchestra.run.vm03.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:23:11.101 INFO:teuthology.orchestra.run.vm03.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:23:11.101 INFO:teuthology.orchestra.run.vm03.stdout:*79.133.44.136 .MBGh. 1 u 14 64 377 20.486 +3.684 3.027 2026-03-06T22:23:11.101 INFO:teuthology.orchestra.run.vm03.stdout:+139-144-71-56.i 80.192.165.246 2 u 10 64 377 22.436 -1.354 3.419 2026-03-06T22:23:11.101 INFO:teuthology.orchestra.run.vm03.stdout:-mail.anyvm.tech 46.48.49.211 3 u 14 64 377 23.495 +3.545 3.338 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm03.stdout:#y.ns.gin.ntt.ne 129.250.35.222 2 u 9 64 377 20.407 -2.281 5.108 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm03.stdout:-ntp2.wup-de.hos 237.17.204.95 2 u 10 64 377 30.968 +4.655 2.915 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm03.stdout:-79.133.44.142 .MBGh. 1 u 10 64 377 20.505 +4.583 3.466 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm03.stdout:#byggvir.de 130.149.17.21 2 u 10 64 377 28.776 -0.206 5.204 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm03.stdout:#82.165.178.31 33.40.230.73 2 u 14 64 377 29.037 -1.788 4.364 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm03.stdout:-139-162-156-95. 82.43.52.28 2 u 7 64 377 22.651 -6.076 5.191 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm03.stdout:-static.215.156. 35.73.197.144 2 u 8 64 377 23.503 +4.694 3.140 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm03.stdout:+cp.hypermediaa. 189.97.54.122 2 u 10 64 377 25.031 +4.343 2.911 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm03.stdout:#185.125.190.56 194.121.207.249 2 u 24 64 377 34.397 -0.162 2.581 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm03.stdout:-172-236-195-26. 233.72.92.146 3 u 5 64 377 23.583 +2.842 2.666 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm03.stdout:#ntp1.wtnet.de 10.129.9.96 2 u 12 64 377 30.809 +1.024 4.548 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm03.stdout:#185.125.190.58 145.238.80.80 2 u 22 64 377 35.790 -1.987 4.653 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm03.stdout:+time2.sebhostin 127.65.222.189 2 u 12 64 377 28.815 +0.540 4.185 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm03.stdout:#185.125.190.57 194.121.207.249 2 u 21 64 377 34.424 +0.958 2.483 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout: remote refid st t when poll reach delay offset jitter 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout:============================================================================== 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout:+79.133.44.136 .MBGh. 1 u 22 64 377 20.560 +0.065 0.106 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout:-mail.anyvm.tech 46.48.49.211 3 u 14 64 377 23.561 -0.638 1.086 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout:+server1a.meinbe 237.17.204.95 2 u 15 64 377 24.998 +0.094 0.036 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout:+cp.hypermediaa. 189.97.54.122 2 u 11 64 377 25.007 -0.184 0.278 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout:#byggvir.de 130.149.17.21 2 u 11 64 377 28.607 +1.282 0.056 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout:#y.ns.gin.ntt.ne 129.250.35.222 2 u 11 64 377 20.344 -0.901 0.029 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout:-172-236-195-26. 233.72.92.146 3 u 8 64 377 25.195 -1.917 0.486 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout:-ntp1.wtnet.de 10.129.9.96 2 u 11 64 377 30.791 +1.481 1.126 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout:#139-162-156-95. 82.43.52.28 2 u 12 64 377 22.475 -5.175 0.324 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout:#82.165.178.31 33.40.230.73 2 u 13 64 377 29.019 -1.530 1.088 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout:-ntp2.wup-de.hos 237.17.204.95 2 u 14 64 377 33.713 +1.445 0.552 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout:#185.125.190.56 194.121.207.249 2 u 27 64 377 35.824 -0.997 0.240 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout:*79.133.44.142 .MBGh. 1 u 12 64 377 20.507 +0.097 1.132 2026-03-06T22:23:11.102 INFO:teuthology.orchestra.run.vm08.stdout:+static.215.156. 35.73.197.144 2 u 15 64 377 23.547 +0.190 1.210 2026-03-06T22:23:11.103 INFO:teuthology.orchestra.run.vm08.stdout:#alphyn.canonica 132.163.96.1 2 u 24 64 377 101.424 -4.041 2.462 2026-03-06T22:23:11.103 INFO:teuthology.orchestra.run.vm08.stdout:+time2.sebhostin 127.65.222.189 2 u 8 64 377 28.972 +0.900 0.249 2026-03-06T22:23:11.103 INFO:teuthology.orchestra.run.vm08.stdout:#185.125.190.57 194.121.207.249 2 u 21 64 377 47.401 -7.324 1.165 2026-03-06T22:23:11.103 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-06T22:23:11.105 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-06T22:23:11.105 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-06T22:23:11.107 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-06T22:23:11.110 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-06T22:23:11.112 INFO:teuthology.task.internal:Duration was 646.066648 seconds 2026-03-06T22:23:11.112 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-06T22:23:11.114 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-06T22:23:11.114 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-06T22:23:11.115 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-06T22:23:11.141 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-06T22:23:11.141 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm03.local 2026-03-06T22:23:11.141 DEBUG:teuthology.orchestra.run.vm03:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-06T22:23:11.197 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm08.local 2026-03-06T22:23:11.198 DEBUG:teuthology.orchestra.run.vm08:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-06T22:23:11.212 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-06T22:23:11.212 DEBUG:teuthology.orchestra.run.vm03:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T22:23:11.240 DEBUG:teuthology.orchestra.run.vm08:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T22:23:11.298 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-06T22:23:11.298 DEBUG:teuthology.orchestra.run.vm03:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T22:23:11.299 DEBUG:teuthology.orchestra.run.vm08:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T22:23:11.305 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T22:23:11.305 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T22:23:11.305 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T22:23:11.305 INFO:teuthology.orchestra.run.vm03.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-06T22:23:11.306 INFO:teuthology.orchestra.run.vm03.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-06T22:23:11.317 INFO:teuthology.orchestra.run.vm03.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 89.7% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-06T22:23:11.348 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T22:23:11.349 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T22:23:11.349 INFO:teuthology.orchestra.run.vm08.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-06T22:23:11.349 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T22:23:11.350 INFO:teuthology.orchestra.run.vm08.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: /home/ubuntu/cephtest/archive/syslog/journalctl.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-06T22:23:11.357 INFO:teuthology.orchestra.run.vm08.stderr: 90.4% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-06T22:23:11.359 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-06T22:23:11.361 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-06T22:23:11.361 DEBUG:teuthology.orchestra.run.vm03:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-06T22:23:11.369 DEBUG:teuthology.orchestra.run.vm08:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-06T22:23:11.412 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-06T22:23:11.415 DEBUG:teuthology.orchestra.run.vm03:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-06T22:23:11.416 DEBUG:teuthology.orchestra.run.vm08:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-06T22:23:11.421 INFO:teuthology.orchestra.run.vm03.stdout:kernel.core_pattern = core 2026-03-06T22:23:11.459 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern = core 2026-03-06T22:23:11.466 DEBUG:teuthology.orchestra.run.vm03:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-06T22:23:11.473 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T22:23:11.473 DEBUG:teuthology.orchestra.run.vm08:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-06T22:23:11.511 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T22:23:11.511 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-06T22:23:11.513 INFO:teuthology.task.internal:Transferring archived files... 2026-03-06T22:23:11.514 DEBUG:teuthology.misc:Transferring archived files from vm03:/home/ubuntu/cephtest/archive to /archive/irq0-2026-03-06_20:21:59-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/382/remote/vm03 2026-03-06T22:23:11.514 DEBUG:teuthology.orchestra.run.vm03:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-06T22:23:11.523 DEBUG:teuthology.misc:Transferring archived files from vm08:/home/ubuntu/cephtest/archive to /archive/irq0-2026-03-06_20:21:59-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/382/remote/vm08 2026-03-06T22:23:11.523 DEBUG:teuthology.orchestra.run.vm08:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-06T22:23:11.562 INFO:teuthology.task.internal:Removing archive directory... 2026-03-06T22:23:11.562 DEBUG:teuthology.orchestra.run.vm03:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-06T22:23:11.568 DEBUG:teuthology.orchestra.run.vm08:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-06T22:23:11.608 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-06T22:23:11.611 INFO:teuthology.task.internal:Not uploading archives. 2026-03-06T22:23:11.611 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-06T22:23:11.613 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-06T22:23:11.613 DEBUG:teuthology.orchestra.run.vm03:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-06T22:23:11.614 DEBUG:teuthology.orchestra.run.vm08:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-06T22:23:11.617 INFO:teuthology.orchestra.run.vm03.stdout: 258077 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 6 22:23 /home/ubuntu/cephtest 2026-03-06T22:23:11.652 INFO:teuthology.orchestra.run.vm08.stdout: 258077 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 6 22:23 /home/ubuntu/cephtest 2026-03-06T22:23:11.653 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-06T22:23:11.659 INFO:teuthology.run:Summary data: description: orch:cephadm:smoke-roleless/{0-distro/ubuntu_22.04 1-start 2-services/nfs-haproxy-proto 3-final} duration: 646.066647529602 owner: irq0 success: true 2026-03-06T22:23:11.659 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-06T22:23:11.682 INFO:teuthology.run:pass