2026-03-10T10:08:36.319 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-10T10:08:36.323 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-10T10:08:36.340 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/996 branch: squid description: orch/cephadm/workunits/{0-distro/ubuntu_22.04 agent/on mon_election/connectivity task/test_set_mon_crush_locations} email: null first_in_suite: false flavor: default job_id: '996' ktype: distro last_in_suite: false machine_type: vps name: kyr-2026-03-10_01:00:38-orch-squid-none-default-vps no_nested_subset: false os_type: ubuntu os_version: '22.04' overrides: admin_socket: branch: squid ansible.cephlab: branch: main skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: UTC ceph: conf: global: mon election default strategy: 3 mgr: debug mgr: 20 debug ms: 1 mgr/cephadm/use_agent: true mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - MON_DOWN - POOL_APP_NOT_ENABLED - mon down - mons down - out of quorum - CEPHADM_FAILED_DAEMON log-only-match: - CEPHADM_ sha1: e911bdebe5c8faa3800735d1568fcdca65db60df ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} install: ceph: flavor: default sha1: e911bdebe5c8faa3800735d1568fcdca65db60df extra_system_packages: deb: - python3-xmltodict - python3-jmespath rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - python3-jmespath workunit: branch: tt-squid sha1: 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - osd.0 - mon.a - mgr.a - - host.b - osd.1 - mon.b - mgr.b - - host.c - osd.2 - mon.c seed: 8043 sha1: e911bdebe5c8faa3800735d1568fcdca65db60df sleep_before_teardown: 0 subset: 1/64 suite: orch suite_branch: tt-squid suite_path: /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b targets: vm00.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBDoiBnDC2KsrkouYtCe/7TK8T1Wb4FpD8RAunLxxR3PmzsxJ5JSUTejAteZJiwS4/MaXHbHUWdkW7DvF/8ljS6Y= vm03.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLDWjcJpit1wPe3SRqYNCI8n2POpfGcsVbwPX+7HDaBNOglqfZgPMBWKzhogeT18PXGTFXCimXl4eoulGZ40jNo= vm09.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKjKgMHC0aROjMvIk9IiYolJCD0sRCVcBlFIl5YUz8+o7m4PBD4nzD1bUY9AhkHhD1xC51QPuC5goIkxzy3nJvM= tasks: - install: null - cephadm: null - cephadm.apply: specs: - placement: count: 3 service_id: foo service_type: mon spec: crush_locations: host.a: - datacenter=a host.b: - datacenter=b - rack=2 host.c: - datacenter=a - rack=3 - cephadm.shell: host.a: - "set -ex\n# since we don't know the real hostnames before the test, the next\n\ # bit is in order to replace the fake hostnames \"host.a/b/c\" with\n# the actual\ \ names cephadm knows the host by within the mon spec\nceph orch host ls --format\ \ json | jq -r '.[] | .hostname' > realnames\necho $'host.a\\nhost.b\\nhost.c'\ \ > fakenames\necho $'a\\nb\\nc' > mon_ids\necho $'{datacenter=a}\\n{datacenter=b,rack=2}\\\ n{datacenter=a,rack=3}' > crush_locs\nceph orch ls --service-name mon --export\ \ > mon.yaml\nMONSPEC=`cat mon.yaml`\necho \"$MONSPEC\"\nwhile read realname\ \ <&3 && read fakename <&4; do\n MONSPEC=\"${MONSPEC//$fakename/$realname}\"\ \ndone 3 mon.yaml\ncat mon.yaml\n\ # now the spec should have the real hostnames, so let's re-apply\nceph orch\ \ apply -i mon.yaml\nsleep 90\nceph orch ps --refresh\nceph orch ls --service-name\ \ mon --export > mon.yaml; ceph orch apply -i mon.yaml\nsleep 90\nceph mon dump\n\ ceph mon dump --format json\n# verify all the crush locations got set from \"\ ceph mon dump\" output\nwhile read monid <&3 && read crushloc <&4; do\n ceph\ \ mon dump --format json | jq --arg monid \"$monid\" --arg crushloc \"$crushloc\"\ \ -e '.mons | .[] | select(.name == $monid) | .crush_location == $crushloc'\n\ done 3, func=.kill_console_loggers at 0x7fe57e5b3e20>, signals=[15]) 2026-03-10T10:08:36.978 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-10T10:08:36.978 INFO:teuthology.task.internal:Opening connections... 2026-03-10T10:08:36.978 DEBUG:teuthology.task.internal:connecting to ubuntu@vm00.local 2026-03-10T10:08:36.979 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm00.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T10:08:37.037 DEBUG:teuthology.task.internal:connecting to ubuntu@vm03.local 2026-03-10T10:08:37.038 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm03.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T10:08:37.098 DEBUG:teuthology.task.internal:connecting to ubuntu@vm09.local 2026-03-10T10:08:37.098 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm09.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T10:08:37.158 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-10T10:08:37.159 DEBUG:teuthology.orchestra.run.vm00:> uname -m 2026-03-10T10:08:37.162 INFO:teuthology.orchestra.run.vm00.stdout:x86_64 2026-03-10T10:08:37.163 DEBUG:teuthology.orchestra.run.vm00:> cat /etc/os-release 2026-03-10T10:08:37.207 INFO:teuthology.orchestra.run.vm00.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-10T10:08:37.207 INFO:teuthology.orchestra.run.vm00.stdout:NAME="Ubuntu" 2026-03-10T10:08:37.207 INFO:teuthology.orchestra.run.vm00.stdout:VERSION_ID="22.04" 2026-03-10T10:08:37.207 INFO:teuthology.orchestra.run.vm00.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-10T10:08:37.207 INFO:teuthology.orchestra.run.vm00.stdout:VERSION_CODENAME=jammy 2026-03-10T10:08:37.207 INFO:teuthology.orchestra.run.vm00.stdout:ID=ubuntu 2026-03-10T10:08:37.207 INFO:teuthology.orchestra.run.vm00.stdout:ID_LIKE=debian 2026-03-10T10:08:37.207 INFO:teuthology.orchestra.run.vm00.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-10T10:08:37.207 INFO:teuthology.orchestra.run.vm00.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-10T10:08:37.207 INFO:teuthology.orchestra.run.vm00.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-10T10:08:37.207 INFO:teuthology.orchestra.run.vm00.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-10T10:08:37.207 INFO:teuthology.orchestra.run.vm00.stdout:UBUNTU_CODENAME=jammy 2026-03-10T10:08:37.207 INFO:teuthology.lock.ops:Updating vm00.local on lock server 2026-03-10T10:08:37.211 DEBUG:teuthology.orchestra.run.vm03:> uname -m 2026-03-10T10:08:37.214 INFO:teuthology.orchestra.run.vm03.stdout:x86_64 2026-03-10T10:08:37.214 DEBUG:teuthology.orchestra.run.vm03:> cat /etc/os-release 2026-03-10T10:08:37.258 INFO:teuthology.orchestra.run.vm03.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-10T10:08:37.258 INFO:teuthology.orchestra.run.vm03.stdout:NAME="Ubuntu" 2026-03-10T10:08:37.258 INFO:teuthology.orchestra.run.vm03.stdout:VERSION_ID="22.04" 2026-03-10T10:08:37.258 INFO:teuthology.orchestra.run.vm03.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-10T10:08:37.258 INFO:teuthology.orchestra.run.vm03.stdout:VERSION_CODENAME=jammy 2026-03-10T10:08:37.258 INFO:teuthology.orchestra.run.vm03.stdout:ID=ubuntu 2026-03-10T10:08:37.258 INFO:teuthology.orchestra.run.vm03.stdout:ID_LIKE=debian 2026-03-10T10:08:37.258 INFO:teuthology.orchestra.run.vm03.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-10T10:08:37.258 INFO:teuthology.orchestra.run.vm03.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-10T10:08:37.258 INFO:teuthology.orchestra.run.vm03.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-10T10:08:37.258 INFO:teuthology.orchestra.run.vm03.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-10T10:08:37.258 INFO:teuthology.orchestra.run.vm03.stdout:UBUNTU_CODENAME=jammy 2026-03-10T10:08:37.258 INFO:teuthology.lock.ops:Updating vm03.local on lock server 2026-03-10T10:08:37.262 DEBUG:teuthology.orchestra.run.vm09:> uname -m 2026-03-10T10:08:37.264 INFO:teuthology.orchestra.run.vm09.stdout:x86_64 2026-03-10T10:08:37.264 DEBUG:teuthology.orchestra.run.vm09:> cat /etc/os-release 2026-03-10T10:08:37.309 INFO:teuthology.orchestra.run.vm09.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-10T10:08:37.309 INFO:teuthology.orchestra.run.vm09.stdout:NAME="Ubuntu" 2026-03-10T10:08:37.309 INFO:teuthology.orchestra.run.vm09.stdout:VERSION_ID="22.04" 2026-03-10T10:08:37.309 INFO:teuthology.orchestra.run.vm09.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-10T10:08:37.309 INFO:teuthology.orchestra.run.vm09.stdout:VERSION_CODENAME=jammy 2026-03-10T10:08:37.309 INFO:teuthology.orchestra.run.vm09.stdout:ID=ubuntu 2026-03-10T10:08:37.309 INFO:teuthology.orchestra.run.vm09.stdout:ID_LIKE=debian 2026-03-10T10:08:37.309 INFO:teuthology.orchestra.run.vm09.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-10T10:08:37.309 INFO:teuthology.orchestra.run.vm09.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-10T10:08:37.309 INFO:teuthology.orchestra.run.vm09.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-10T10:08:37.309 INFO:teuthology.orchestra.run.vm09.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-10T10:08:37.309 INFO:teuthology.orchestra.run.vm09.stdout:UBUNTU_CODENAME=jammy 2026-03-10T10:08:37.309 INFO:teuthology.lock.ops:Updating vm09.local on lock server 2026-03-10T10:08:37.313 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-10T10:08:37.314 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-10T10:08:37.315 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-10T10:08:37.315 DEBUG:teuthology.orchestra.run.vm00:> test '!' -e /home/ubuntu/cephtest 2026-03-10T10:08:37.316 DEBUG:teuthology.orchestra.run.vm03:> test '!' -e /home/ubuntu/cephtest 2026-03-10T10:08:37.317 DEBUG:teuthology.orchestra.run.vm09:> test '!' -e /home/ubuntu/cephtest 2026-03-10T10:08:37.352 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-10T10:08:37.353 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-10T10:08:37.353 DEBUG:teuthology.orchestra.run.vm00:> test -z $(ls -A /var/lib/ceph) 2026-03-10T10:08:37.361 DEBUG:teuthology.orchestra.run.vm03:> test -z $(ls -A /var/lib/ceph) 2026-03-10T10:08:37.361 DEBUG:teuthology.orchestra.run.vm09:> test -z $(ls -A /var/lib/ceph) 2026-03-10T10:08:37.362 INFO:teuthology.orchestra.run.vm00.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T10:08:37.363 INFO:teuthology.orchestra.run.vm03.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T10:08:37.397 INFO:teuthology.orchestra.run.vm09.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T10:08:37.397 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-10T10:08:37.404 DEBUG:teuthology.orchestra.run.vm00:> test -e /ceph-qa-ready 2026-03-10T10:08:37.406 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:08:37.631 DEBUG:teuthology.orchestra.run.vm03:> test -e /ceph-qa-ready 2026-03-10T10:08:37.634 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:08:38.047 DEBUG:teuthology.orchestra.run.vm09:> test -e /ceph-qa-ready 2026-03-10T10:08:38.049 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:08:38.266 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-10T10:08:38.267 INFO:teuthology.task.internal:Creating test directory... 2026-03-10T10:08:38.267 DEBUG:teuthology.orchestra.run.vm00:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T10:08:38.268 DEBUG:teuthology.orchestra.run.vm03:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T10:08:38.269 DEBUG:teuthology.orchestra.run.vm09:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T10:08:38.271 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-10T10:08:38.272 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-10T10:08:38.273 INFO:teuthology.task.internal:Creating archive directory... 2026-03-10T10:08:38.274 DEBUG:teuthology.orchestra.run.vm00:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T10:08:38.312 DEBUG:teuthology.orchestra.run.vm03:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T10:08:38.313 DEBUG:teuthology.orchestra.run.vm09:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T10:08:38.317 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-10T10:08:38.318 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-10T10:08:38.318 DEBUG:teuthology.orchestra.run.vm00:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T10:08:38.357 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:08:38.358 DEBUG:teuthology.orchestra.run.vm03:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T10:08:38.359 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:08:38.359 DEBUG:teuthology.orchestra.run.vm09:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T10:08:38.361 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:08:38.361 DEBUG:teuthology.orchestra.run.vm00:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T10:08:38.400 DEBUG:teuthology.orchestra.run.vm03:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T10:08:38.404 DEBUG:teuthology.orchestra.run.vm09:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T10:08:38.406 INFO:teuthology.orchestra.run.vm00.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T10:08:38.411 INFO:teuthology.orchestra.run.vm03.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T10:08:38.411 INFO:teuthology.orchestra.run.vm00.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T10:08:38.411 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T10:08:38.415 INFO:teuthology.orchestra.run.vm03.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T10:08:38.415 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T10:08:38.416 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-10T10:08:38.417 INFO:teuthology.task.internal:Configuring sudo... 2026-03-10T10:08:38.417 DEBUG:teuthology.orchestra.run.vm00:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T10:08:38.456 DEBUG:teuthology.orchestra.run.vm03:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T10:08:38.460 DEBUG:teuthology.orchestra.run.vm09:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T10:08:38.466 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-10T10:08:38.468 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-10T10:08:38.468 DEBUG:teuthology.orchestra.run.vm00:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T10:08:38.504 DEBUG:teuthology.orchestra.run.vm03:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T10:08:38.508 DEBUG:teuthology.orchestra.run.vm09:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T10:08:38.512 DEBUG:teuthology.orchestra.run.vm00:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T10:08:38.550 DEBUG:teuthology.orchestra.run.vm00:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T10:08:38.593 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-10T10:08:38.593 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T10:08:38.642 DEBUG:teuthology.orchestra.run.vm03:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T10:08:38.645 DEBUG:teuthology.orchestra.run.vm03:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T10:08:38.689 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-10T10:08:38.689 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T10:08:38.738 DEBUG:teuthology.orchestra.run.vm09:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T10:08:38.740 DEBUG:teuthology.orchestra.run.vm09:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T10:08:38.784 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T10:08:38.784 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T10:08:38.833 DEBUG:teuthology.orchestra.run.vm00:> sudo service rsyslog restart 2026-03-10T10:08:38.834 DEBUG:teuthology.orchestra.run.vm03:> sudo service rsyslog restart 2026-03-10T10:08:38.835 DEBUG:teuthology.orchestra.run.vm09:> sudo service rsyslog restart 2026-03-10T10:08:38.889 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-10T10:08:38.891 INFO:teuthology.task.internal:Starting timer... 2026-03-10T10:08:38.891 INFO:teuthology.run_tasks:Running task pcp... 2026-03-10T10:08:38.893 INFO:teuthology.run_tasks:Running task selinux... 2026-03-10T10:08:38.895 INFO:teuthology.task.selinux:Excluding vm00: VMs are not yet supported 2026-03-10T10:08:38.895 INFO:teuthology.task.selinux:Excluding vm03: VMs are not yet supported 2026-03-10T10:08:38.895 INFO:teuthology.task.selinux:Excluding vm09: VMs are not yet supported 2026-03-10T10:08:38.895 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-10T10:08:38.895 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-10T10:08:38.895 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-10T10:08:38.895 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-10T10:08:38.897 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'UTC'}} 2026-03-10T10:08:38.897 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/ceph/ceph-cm-ansible.git 2026-03-10T10:08:38.899 INFO:teuthology.repo_utils:Fetching github.com_ceph_ceph-cm-ansible_main from origin 2026-03-10T10:08:39.399 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main to origin/main 2026-03-10T10:08:39.405 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-10T10:08:39.405 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "UTC"}' -i /tmp/teuth_ansible_inventorydgnfxb13 --limit vm00.local,vm03.local,vm09.local /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-10T10:10:47.900 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm00.local'), Remote(name='ubuntu@vm03.local'), Remote(name='ubuntu@vm09.local')] 2026-03-10T10:10:47.900 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm00.local' 2026-03-10T10:10:47.900 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm00.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T10:10:47.959 DEBUG:teuthology.orchestra.run.vm00:> true 2026-03-10T10:10:48.176 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm00.local' 2026-03-10T10:10:48.176 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm03.local' 2026-03-10T10:10:48.177 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm03.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T10:10:48.239 DEBUG:teuthology.orchestra.run.vm03:> true 2026-03-10T10:10:48.469 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm03.local' 2026-03-10T10:10:48.469 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm09.local' 2026-03-10T10:10:48.469 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm09.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T10:10:48.532 DEBUG:teuthology.orchestra.run.vm09:> true 2026-03-10T10:10:48.752 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm09.local' 2026-03-10T10:10:48.753 INFO:teuthology.run_tasks:Running task clock... 2026-03-10T10:10:48.755 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-10T10:10:48.755 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T10:10:48.755 DEBUG:teuthology.orchestra.run.vm00:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T10:10:48.756 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T10:10:48.756 DEBUG:teuthology.orchestra.run.vm03:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T10:10:48.757 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T10:10:48.757 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T10:10:48.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-10T10:10:48.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: Command line: ntpd -gq 2026-03-10T10:10:48.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: ---------------------------------------------------- 2026-03-10T10:10:48.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: ntp-4 is maintained by Network Time Foundation, 2026-03-10T10:10:48.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-10T10:10:48.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: corporation. Support and training for ntp-4 are 2026-03-10T10:10:48.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: available at https://www.nwtime.org/support 2026-03-10T10:10:48.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: ---------------------------------------------------- 2026-03-10T10:10:48.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: proto: precision = 0.029 usec (-25) 2026-03-10T10:10:48.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: basedate set to 2022-02-04 2026-03-10T10:10:48.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: gps base set to 2022-02-06 (week 2196) 2026-03-10T10:10:48.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-10T10:10:48.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-10T10:10:48.771 INFO:teuthology.orchestra.run.vm00.stderr:10 Mar 10:10:48 ntpd[16094]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 73 days ago 2026-03-10T10:10:48.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: Listen and drop on 0 v6wildcard [::]:123 2026-03-10T10:10:48.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-10T10:10:48.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: Listen normally on 2 lo 127.0.0.1:123 2026-03-10T10:10:48.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: Listen normally on 3 ens3 192.168.123.100:123 2026-03-10T10:10:48.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: Listen normally on 4 lo [::1]:123 2026-03-10T10:10:48.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:0%2]:123 2026-03-10T10:10:48.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:48 ntpd[16094]: Listening on routing socket on fd #22 for interface updates 2026-03-10T10:10:48.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-10T10:10:48.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: Command line: ntpd -gq 2026-03-10T10:10:48.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: ---------------------------------------------------- 2026-03-10T10:10:48.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: ntp-4 is maintained by Network Time Foundation, 2026-03-10T10:10:48.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-10T10:10:48.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: corporation. Support and training for ntp-4 are 2026-03-10T10:10:48.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: available at https://www.nwtime.org/support 2026-03-10T10:10:48.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: ---------------------------------------------------- 2026-03-10T10:10:48.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: proto: precision = 0.030 usec (-25) 2026-03-10T10:10:48.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: basedate set to 2022-02-04 2026-03-10T10:10:48.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: gps base set to 2022-02-06 (week 2196) 2026-03-10T10:10:48.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-10T10:10:48.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-10T10:10:48.773 INFO:teuthology.orchestra.run.vm03.stderr:10 Mar 10:10:48 ntpd[16099]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 73 days ago 2026-03-10T10:10:48.774 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: Listen and drop on 0 v6wildcard [::]:123 2026-03-10T10:10:48.774 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-10T10:10:48.774 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: Listen normally on 2 lo 127.0.0.1:123 2026-03-10T10:10:48.774 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: Listen normally on 3 ens3 192.168.123.103:123 2026-03-10T10:10:48.774 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: Listen normally on 4 lo [::1]:123 2026-03-10T10:10:48.774 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:3%2]:123 2026-03-10T10:10:48.774 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:48 ntpd[16099]: Listening on routing socket on fd #22 for interface updates 2026-03-10T10:10:48.808 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-10T10:10:48.808 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: Command line: ntpd -gq 2026-03-10T10:10:48.808 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: ---------------------------------------------------- 2026-03-10T10:10:48.808 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: ntp-4 is maintained by Network Time Foundation, 2026-03-10T10:10:48.808 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-10T10:10:48.808 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: corporation. Support and training for ntp-4 are 2026-03-10T10:10:48.808 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: available at https://www.nwtime.org/support 2026-03-10T10:10:48.808 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: ---------------------------------------------------- 2026-03-10T10:10:48.808 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: proto: precision = 0.029 usec (-25) 2026-03-10T10:10:48.808 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: basedate set to 2022-02-04 2026-03-10T10:10:48.808 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: gps base set to 2022-02-06 (week 2196) 2026-03-10T10:10:48.808 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-10T10:10:48.808 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-10T10:10:48.808 INFO:teuthology.orchestra.run.vm09.stderr:10 Mar 10:10:48 ntpd[16094]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 73 days ago 2026-03-10T10:10:48.809 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: Listen and drop on 0 v6wildcard [::]:123 2026-03-10T10:10:48.809 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-10T10:10:48.809 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: Listen normally on 2 lo 127.0.0.1:123 2026-03-10T10:10:48.809 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: Listen normally on 3 ens3 192.168.123.109:123 2026-03-10T10:10:48.809 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: Listen normally on 4 lo [::1]:123 2026-03-10T10:10:48.809 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:9%2]:123 2026-03-10T10:10:48.809 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:48 ntpd[16094]: Listening on routing socket on fd #22 for interface updates 2026-03-10T10:10:49.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:49 ntpd[16094]: Soliciting pool server 5.189.151.39 2026-03-10T10:10:49.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:49 ntpd[16099]: Soliciting pool server 5.189.151.39 2026-03-10T10:10:49.808 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:49 ntpd[16094]: Soliciting pool server 31.209.85.242 2026-03-10T10:10:50.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:50 ntpd[16094]: Soliciting pool server 144.76.66.156 2026-03-10T10:10:50.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:50 ntpd[16094]: Soliciting pool server 78.46.56.170 2026-03-10T10:10:50.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:50 ntpd[16099]: Soliciting pool server 144.76.66.156 2026-03-10T10:10:50.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:50 ntpd[16099]: Soliciting pool server 78.46.56.170 2026-03-10T10:10:50.807 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:50 ntpd[16094]: Soliciting pool server 5.189.151.39 2026-03-10T10:10:50.808 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:50 ntpd[16094]: Soliciting pool server 176.9.42.91 2026-03-10T10:10:51.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:51 ntpd[16094]: Soliciting pool server 77.42.16.222 2026-03-10T10:10:51.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:51 ntpd[16094]: Soliciting pool server 139.162.152.20 2026-03-10T10:10:51.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:51 ntpd[16094]: Soliciting pool server 185.11.138.90 2026-03-10T10:10:51.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:51 ntpd[16099]: Soliciting pool server 77.42.16.222 2026-03-10T10:10:51.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:51 ntpd[16099]: Soliciting pool server 139.162.152.20 2026-03-10T10:10:51.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:51 ntpd[16099]: Soliciting pool server 185.11.138.90 2026-03-10T10:10:51.807 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:51 ntpd[16094]: Soliciting pool server 78.46.56.170 2026-03-10T10:10:51.807 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:51 ntpd[16094]: Soliciting pool server 144.76.66.156 2026-03-10T10:10:51.807 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:51 ntpd[16094]: Soliciting pool server 185.252.140.125 2026-03-10T10:10:52.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:52 ntpd[16094]: Soliciting pool server 104.167.24.26 2026-03-10T10:10:52.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:52 ntpd[16094]: Soliciting pool server 78.47.249.55 2026-03-10T10:10:52.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:52 ntpd[16094]: Soliciting pool server 31.209.85.242 2026-03-10T10:10:52.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:52 ntpd[16094]: Soliciting pool server 129.70.132.35 2026-03-10T10:10:52.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:52 ntpd[16099]: Soliciting pool server 104.167.24.26 2026-03-10T10:10:52.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:52 ntpd[16099]: Soliciting pool server 78.47.249.55 2026-03-10T10:10:52.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:52 ntpd[16099]: Soliciting pool server 31.209.85.242 2026-03-10T10:10:52.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:52 ntpd[16099]: Soliciting pool server 129.70.132.35 2026-03-10T10:10:52.806 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:52 ntpd[16094]: Soliciting pool server 185.11.138.90 2026-03-10T10:10:52.807 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:52 ntpd[16094]: Soliciting pool server 77.42.16.222 2026-03-10T10:10:52.807 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:52 ntpd[16094]: Soliciting pool server 139.162.152.20 2026-03-10T10:10:52.807 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:52 ntpd[16094]: Soliciting pool server 141.144.241.16 2026-03-10T10:10:53.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:53 ntpd[16094]: Soliciting pool server 5.45.97.204 2026-03-10T10:10:53.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:53 ntpd[16094]: Soliciting pool server 79.133.44.142 2026-03-10T10:10:53.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:53 ntpd[16094]: Soliciting pool server 176.9.42.91 2026-03-10T10:10:53.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:53 ntpd[16094]: Soliciting pool server 185.125.190.57 2026-03-10T10:10:53.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:53 ntpd[16099]: Soliciting pool server 5.45.97.204 2026-03-10T10:10:53.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:53 ntpd[16099]: Soliciting pool server 79.133.44.142 2026-03-10T10:10:53.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:53 ntpd[16099]: Soliciting pool server 176.9.42.91 2026-03-10T10:10:53.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:53 ntpd[16099]: Soliciting pool server 185.125.190.57 2026-03-10T10:10:53.806 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:53 ntpd[16094]: Soliciting pool server 129.70.132.35 2026-03-10T10:10:53.806 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:53 ntpd[16094]: Soliciting pool server 104.167.24.26 2026-03-10T10:10:53.806 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:53 ntpd[16094]: Soliciting pool server 78.47.249.55 2026-03-10T10:10:53.807 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:53 ntpd[16094]: Soliciting pool server 91.189.91.157 2026-03-10T10:10:54.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:54 ntpd[16094]: Soliciting pool server 185.125.190.56 2026-03-10T10:10:54.771 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:54 ntpd[16094]: Soliciting pool server 62.108.36.235 2026-03-10T10:10:54.772 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:54 ntpd[16094]: Soliciting pool server 185.252.140.125 2026-03-10T10:10:54.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:54 ntpd[16099]: Soliciting pool server 185.125.190.56 2026-03-10T10:10:54.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:54 ntpd[16099]: Soliciting pool server 62.108.36.235 2026-03-10T10:10:54.773 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:54 ntpd[16099]: Soliciting pool server 185.252.140.125 2026-03-10T10:10:54.805 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:54 ntpd[16094]: Soliciting pool server 185.125.190.57 2026-03-10T10:10:54.805 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:54 ntpd[16094]: Soliciting pool server 5.45.97.204 2026-03-10T10:10:54.806 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:54 ntpd[16094]: Soliciting pool server 79.133.44.142 2026-03-10T10:10:56.835 INFO:teuthology.orchestra.run.vm09.stdout:10 Mar 10:10:56 ntpd[16094]: ntpd: time slew +0.014730 s 2026-03-10T10:10:56.835 INFO:teuthology.orchestra.run.vm09.stdout:ntpd: time slew +0.014730s 2026-03-10T10:10:56.856 INFO:teuthology.orchestra.run.vm09.stdout: remote refid st t when poll reach delay offset jitter 2026-03-10T10:10:56.857 INFO:teuthology.orchestra.run.vm09.stdout:============================================================================== 2026-03-10T10:10:56.857 INFO:teuthology.orchestra.run.vm09.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:10:56.857 INFO:teuthology.orchestra.run.vm09.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:10:56.857 INFO:teuthology.orchestra.run.vm09.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:10:56.857 INFO:teuthology.orchestra.run.vm09.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:10:56.857 INFO:teuthology.orchestra.run.vm09.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:10:57.798 INFO:teuthology.orchestra.run.vm03.stdout:10 Mar 10:10:57 ntpd[16099]: ntpd: time slew +0.000363 s 2026-03-10T10:10:57.798 INFO:teuthology.orchestra.run.vm03.stdout:ntpd: time slew +0.000363s 2026-03-10T10:10:57.806 INFO:teuthology.orchestra.run.vm00.stdout:10 Mar 10:10:57 ntpd[16094]: ntpd: time slew -0.001086 s 2026-03-10T10:10:57.806 INFO:teuthology.orchestra.run.vm00.stdout:ntpd: time slew -0.001086s 2026-03-10T10:10:57.819 INFO:teuthology.orchestra.run.vm03.stdout: remote refid st t when poll reach delay offset jitter 2026-03-10T10:10:57.819 INFO:teuthology.orchestra.run.vm03.stdout:============================================================================== 2026-03-10T10:10:57.819 INFO:teuthology.orchestra.run.vm03.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:10:57.819 INFO:teuthology.orchestra.run.vm03.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:10:57.819 INFO:teuthology.orchestra.run.vm03.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:10:57.819 INFO:teuthology.orchestra.run.vm03.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:10:57.819 INFO:teuthology.orchestra.run.vm03.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:10:57.828 INFO:teuthology.orchestra.run.vm00.stdout: remote refid st t when poll reach delay offset jitter 2026-03-10T10:10:57.828 INFO:teuthology.orchestra.run.vm00.stdout:============================================================================== 2026-03-10T10:10:57.828 INFO:teuthology.orchestra.run.vm00.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:10:57.828 INFO:teuthology.orchestra.run.vm00.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:10:57.828 INFO:teuthology.orchestra.run.vm00.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:10:57.828 INFO:teuthology.orchestra.run.vm00.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:10:57.828 INFO:teuthology.orchestra.run.vm00.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:10:57.829 INFO:teuthology.run_tasks:Running task install... 2026-03-10T10:10:57.831 DEBUG:teuthology.task.install:project ceph 2026-03-10T10:10:57.831 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'}, 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-10T10:10:57.831 DEBUG:teuthology.task.install:config {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-10T10:10:57.831 INFO:teuthology.task.install:Using flavor: default 2026-03-10T10:10:57.833 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-03-10T10:10:57.833 INFO:teuthology.task.install:extra packages: [] 2026-03-10T10:10:57.833 DEBUG:teuthology.orchestra.run.vm00:> sudo apt-key list | grep Ceph 2026-03-10T10:10:57.833 DEBUG:teuthology.orchestra.run.vm03:> sudo apt-key list | grep Ceph 2026-03-10T10:10:57.833 DEBUG:teuthology.orchestra.run.vm09:> sudo apt-key list | grep Ceph 2026-03-10T10:10:57.869 INFO:teuthology.orchestra.run.vm09.stderr:Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). 2026-03-10T10:10:57.886 INFO:teuthology.orchestra.run.vm09.stdout:uid [ unknown] Ceph automated package build (Ceph automated package build) 2026-03-10T10:10:57.886 INFO:teuthology.orchestra.run.vm09.stdout:uid [ unknown] Ceph.com (release key) 2026-03-10T10:10:57.887 INFO:teuthology.task.install.deb:Installing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on remote deb x86_64 2026-03-10T10:10:57.887 INFO:teuthology.task.install.deb:Installing system (non-project) packages: python3-xmltodict, python3-jmespath on remote deb x86_64 2026-03-10T10:10:57.887 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T10:10:57.947 INFO:teuthology.orchestra.run.vm00.stderr:Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). 2026-03-10T10:10:57.947 INFO:teuthology.orchestra.run.vm03.stderr:Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). 2026-03-10T10:10:57.947 INFO:teuthology.orchestra.run.vm00.stdout:uid [ unknown] Ceph automated package build (Ceph automated package build) 2026-03-10T10:10:57.947 INFO:teuthology.orchestra.run.vm00.stdout:uid [ unknown] Ceph.com (release key) 2026-03-10T10:10:57.947 INFO:teuthology.orchestra.run.vm03.stdout:uid [ unknown] Ceph automated package build (Ceph automated package build) 2026-03-10T10:10:57.947 INFO:teuthology.orchestra.run.vm03.stdout:uid [ unknown] Ceph.com (release key) 2026-03-10T10:10:57.947 INFO:teuthology.task.install.deb:Installing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on remote deb x86_64 2026-03-10T10:10:57.947 INFO:teuthology.task.install.deb:Installing system (non-project) packages: python3-xmltodict, python3-jmespath on remote deb x86_64 2026-03-10T10:10:57.947 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T10:10:57.948 INFO:teuthology.task.install.deb:Installing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on remote deb x86_64 2026-03-10T10:10:57.948 INFO:teuthology.task.install.deb:Installing system (non-project) packages: python3-xmltodict, python3-jmespath on remote deb x86_64 2026-03-10T10:10:57.948 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T10:10:58.549 INFO:teuthology.task.install.deb:Pulling from https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default/ 2026-03-10T10:10:58.549 INFO:teuthology.task.install.deb:Package version is 19.2.3-678-ge911bdeb-1jammy 2026-03-10T10:10:58.611 INFO:teuthology.task.install.deb:Pulling from https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default/ 2026-03-10T10:10:58.611 INFO:teuthology.task.install.deb:Package version is 19.2.3-678-ge911bdeb-1jammy 2026-03-10T10:10:58.700 INFO:teuthology.task.install.deb:Pulling from https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default/ 2026-03-10T10:10:58.700 INFO:teuthology.task.install.deb:Package version is 19.2.3-678-ge911bdeb-1jammy 2026-03-10T10:10:59.191 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T10:10:59.192 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/apt/sources.list.d/ceph.list 2026-03-10T10:10:59.199 DEBUG:teuthology.orchestra.run.vm09:> sudo apt-get update 2026-03-10T10:10:59.299 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-10T10:10:59.299 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/etc/apt/sources.list.d/ceph.list 2026-03-10T10:10:59.306 DEBUG:teuthology.orchestra.run.vm00:> sudo apt-get update 2026-03-10T10:10:59.327 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-10T10:10:59.327 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/etc/apt/sources.list.d/ceph.list 2026-03-10T10:10:59.335 DEBUG:teuthology.orchestra.run.vm03:> sudo apt-get update 2026-03-10T10:10:59.371 INFO:teuthology.orchestra.run.vm09.stdout:Hit:1 https://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-10T10:10:59.506 INFO:teuthology.orchestra.run.vm00.stdout:Hit:1 https://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-10T10:10:59.776 INFO:teuthology.orchestra.run.vm03.stdout:Hit:1 https://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-10T10:10:59.779 INFO:teuthology.orchestra.run.vm09.stdout:Hit:2 https://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-10T10:10:59.846 INFO:teuthology.orchestra.run.vm09.stdout:Ign:3 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy InRelease 2026-03-10T10:10:59.878 INFO:teuthology.orchestra.run.vm03.stdout:Hit:2 https://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-10T10:10:59.882 INFO:teuthology.orchestra.run.vm09.stdout:Hit:4 https://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-10T10:10:59.977 INFO:teuthology.orchestra.run.vm03.stdout:Hit:3 https://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-10T10:10:59.984 INFO:teuthology.orchestra.run.vm09.stdout:Get:5 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy Release [7662 B] 2026-03-10T10:10:59.985 INFO:teuthology.orchestra.run.vm09.stdout:Hit:6 https://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-10T10:11:00.008 INFO:teuthology.orchestra.run.vm00.stdout:Ign:2 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy InRelease 2026-03-10T10:11:00.021 INFO:teuthology.orchestra.run.vm03.stdout:Ign:4 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy InRelease 2026-03-10T10:11:00.077 INFO:teuthology.orchestra.run.vm03.stdout:Hit:5 https://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-10T10:11:00.124 INFO:teuthology.orchestra.run.vm09.stdout:Ign:7 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy Release.gpg 2026-03-10T10:11:00.129 INFO:teuthology.orchestra.run.vm00.stdout:Hit:3 https://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-10T10:11:00.160 INFO:teuthology.orchestra.run.vm00.stdout:Get:4 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy Release [7662 B] 2026-03-10T10:11:00.172 INFO:teuthology.orchestra.run.vm03.stdout:Get:6 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy Release [7662 B] 2026-03-10T10:11:00.232 INFO:teuthology.orchestra.run.vm00.stdout:Hit:5 https://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-10T10:11:00.276 INFO:teuthology.orchestra.run.vm09.stdout:Get:8 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 Packages [18.1 kB] 2026-03-10T10:11:00.323 INFO:teuthology.orchestra.run.vm00.stdout:Ign:6 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy Release.gpg 2026-03-10T10:11:00.333 INFO:teuthology.orchestra.run.vm03.stdout:Ign:7 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy Release.gpg 2026-03-10T10:11:00.334 INFO:teuthology.orchestra.run.vm00.stdout:Hit:7 https://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-10T10:11:00.352 INFO:teuthology.orchestra.run.vm09.stdout:Fetched 25.8 kB in 1s (25.6 kB/s) 2026-03-10T10:11:00.488 INFO:teuthology.orchestra.run.vm00.stdout:Get:8 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 Packages [18.1 kB] 2026-03-10T10:11:00.496 INFO:teuthology.orchestra.run.vm03.stdout:Get:8 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 Packages [18.1 kB] 2026-03-10T10:11:00.562 INFO:teuthology.orchestra.run.vm00.stdout:Fetched 25.8 kB in 1s (23.5 kB/s) 2026-03-10T10:11:00.577 INFO:teuthology.orchestra.run.vm03.stdout:Fetched 25.8 kB in 1s (23.8 kB/s) 2026-03-10T10:11:01.045 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:11:01.058 DEBUG:teuthology.orchestra.run.vm09:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=19.2.3-678-ge911bdeb-1jammy cephadm=19.2.3-678-ge911bdeb-1jammy ceph-mds=19.2.3-678-ge911bdeb-1jammy ceph-mgr=19.2.3-678-ge911bdeb-1jammy ceph-common=19.2.3-678-ge911bdeb-1jammy ceph-fuse=19.2.3-678-ge911bdeb-1jammy ceph-test=19.2.3-678-ge911bdeb-1jammy ceph-volume=19.2.3-678-ge911bdeb-1jammy radosgw=19.2.3-678-ge911bdeb-1jammy python3-rados=19.2.3-678-ge911bdeb-1jammy python3-rgw=19.2.3-678-ge911bdeb-1jammy python3-cephfs=19.2.3-678-ge911bdeb-1jammy python3-rbd=19.2.3-678-ge911bdeb-1jammy libcephfs2=19.2.3-678-ge911bdeb-1jammy libcephfs-dev=19.2.3-678-ge911bdeb-1jammy librados2=19.2.3-678-ge911bdeb-1jammy librbd1=19.2.3-678-ge911bdeb-1jammy rbd-fuse=19.2.3-678-ge911bdeb-1jammy 2026-03-10T10:11:01.092 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:11:01.232 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:11:01.245 DEBUG:teuthology.orchestra.run.vm00:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=19.2.3-678-ge911bdeb-1jammy cephadm=19.2.3-678-ge911bdeb-1jammy ceph-mds=19.2.3-678-ge911bdeb-1jammy ceph-mgr=19.2.3-678-ge911bdeb-1jammy ceph-common=19.2.3-678-ge911bdeb-1jammy ceph-fuse=19.2.3-678-ge911bdeb-1jammy ceph-test=19.2.3-678-ge911bdeb-1jammy ceph-volume=19.2.3-678-ge911bdeb-1jammy radosgw=19.2.3-678-ge911bdeb-1jammy python3-rados=19.2.3-678-ge911bdeb-1jammy python3-rgw=19.2.3-678-ge911bdeb-1jammy python3-cephfs=19.2.3-678-ge911bdeb-1jammy python3-rbd=19.2.3-678-ge911bdeb-1jammy libcephfs2=19.2.3-678-ge911bdeb-1jammy libcephfs-dev=19.2.3-678-ge911bdeb-1jammy librados2=19.2.3-678-ge911bdeb-1jammy librbd1=19.2.3-678-ge911bdeb-1jammy rbd-fuse=19.2.3-678-ge911bdeb-1jammy 2026-03-10T10:11:01.279 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:11:01.281 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:11:01.290 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:11:01.290 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:11:01.292 DEBUG:teuthology.orchestra.run.vm03:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=19.2.3-678-ge911bdeb-1jammy cephadm=19.2.3-678-ge911bdeb-1jammy ceph-mds=19.2.3-678-ge911bdeb-1jammy ceph-mgr=19.2.3-678-ge911bdeb-1jammy ceph-common=19.2.3-678-ge911bdeb-1jammy ceph-fuse=19.2.3-678-ge911bdeb-1jammy ceph-test=19.2.3-678-ge911bdeb-1jammy ceph-volume=19.2.3-678-ge911bdeb-1jammy radosgw=19.2.3-678-ge911bdeb-1jammy python3-rados=19.2.3-678-ge911bdeb-1jammy python3-rgw=19.2.3-678-ge911bdeb-1jammy python3-cephfs=19.2.3-678-ge911bdeb-1jammy python3-rbd=19.2.3-678-ge911bdeb-1jammy libcephfs2=19.2.3-678-ge911bdeb-1jammy libcephfs-dev=19.2.3-678-ge911bdeb-1jammy librados2=19.2.3-678-ge911bdeb-1jammy librbd1=19.2.3-678-ge911bdeb-1jammy rbd-fuse=19.2.3-678-ge911bdeb-1jammy 2026-03-10T10:11:01.324 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:11:01.396 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:11:01.396 INFO:teuthology.orchestra.run.vm09.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T10:11:01.396 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-10T10:11:01.396 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:11:01.396 INFO:teuthology.orchestra.run.vm09.stdout:The following additional packages will be installed: 2026-03-10T10:11:01.396 INFO:teuthology.orchestra.run.vm09.stdout: ceph-base ceph-mgr-cephadm ceph-mgr-dashboard ceph-mgr-diskprediction-local 2026-03-10T10:11:01.396 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-k8sevents ceph-mgr-modules-core ceph-mon ceph-osd jq 2026-03-10T10:11:01.396 INFO:teuthology.orchestra.run.vm09.stdout: libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T10:11:01.396 INFO:teuthology.orchestra.run.vm09.stdout: liboath0 libonig5 libpcre2-16-0 libqt5core5a libqt5dbus5 libqt5network5 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsqlite3-mod-ceph 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout: libthrift-0.16.0 lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout: python3-pastescript python3-pecan python3-pluggy python3-portend 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyinotify python3-pytest python3-repoze.lru 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout: python3-toml python3-waitress python3-wcwidth python3-webob 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket python3-webtest python3-werkzeug python3-zc.lockfile 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout: qttranslations5-l10n smartmontools socat unzip xmlstarlet zip 2026-03-10T10:11:01.397 INFO:teuthology.orchestra.run.vm09.stdout:Suggested packages: 2026-03-10T10:11:01.398 INFO:teuthology.orchestra.run.vm09.stdout: python3-influxdb readline-doc python3-beaker python-mako-doc 2026-03-10T10:11:01.398 INFO:teuthology.orchestra.run.vm09.stdout: python-natsort-doc httpd-wsgi libapache2-mod-python libapache2-mod-scgi 2026-03-10T10:11:01.398 INFO:teuthology.orchestra.run.vm09.stdout: libjs-mochikit python-pecan-doc python-psutil-doc subversion 2026-03-10T10:11:01.398 INFO:teuthology.orchestra.run.vm09.stdout: python-pygments-doc ttf-bitstream-vera python-pyinotify-doc python3-dap 2026-03-10T10:11:01.398 INFO:teuthology.orchestra.run.vm09.stdout: python-sklearn-doc ipython3 python-waitress-doc python-webob-doc 2026-03-10T10:11:01.398 INFO:teuthology.orchestra.run.vm09.stdout: python-webtest-doc python-werkzeug-doc python3-watchdog gsmartcontrol 2026-03-10T10:11:01.398 INFO:teuthology.orchestra.run.vm09.stdout: smart-notifier mailx | mailutils 2026-03-10T10:11:01.398 INFO:teuthology.orchestra.run.vm09.stdout:Recommended packages: 2026-03-10T10:11:01.398 INFO:teuthology.orchestra.run.vm09.stdout: btrfs-tools 2026-03-10T10:11:01.431 INFO:teuthology.orchestra.run.vm09.stdout:The following NEW packages will be installed: 2026-03-10T10:11:01.431 INFO:teuthology.orchestra.run.vm09.stdout: ceph ceph-base ceph-common ceph-fuse ceph-mds ceph-mgr ceph-mgr-cephadm 2026-03-10T10:11:01.431 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-k8sevents 2026-03-10T10:11:01.431 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core ceph-mon ceph-osd ceph-test ceph-volume cephadm jq 2026-03-10T10:11:01.431 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-dev libcephfs2 libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 2026-03-10T10:11:01.431 INFO:teuthology.orchestra.run.vm09.stdout: liblua5.3-dev libnbd0 liboath0 libonig5 libpcre2-16-0 libqt5core5a 2026-03-10T10:11:01.431 INFO:teuthology.orchestra.run.vm09.stdout: libqt5dbus5 libqt5network5 libradosstriper1 librdkafka1 libreadline-dev 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: librgw2 libsqlite3-mod-ceph libthrift-0.16.0 lua-any lua-sec lua-socket 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: lua5.1 luarocks nvme-cli pkg-config python-asyncssh-doc 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse python3-ceph-common python3-cephfs python3-cheroot 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: python3-pastescript python3-pecan python3-pluggy python3-portend 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyinotify python3-pytest python3-rados python3-rbd 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: python3-repoze.lru python3-requests-oauthlib python3-rgw python3-routes 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa python3-simplegeneric python3-simplejson python3-singledispatch 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: python3-sklearn python3-sklearn-lib python3-tempita python3-tempora 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: python3-threadpoolctl python3-toml python3-waitress python3-wcwidth 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc.lockfile qttranslations5-l10n radosgw rbd-fuse smartmontools 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: socat unzip xmlstarlet zip 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be upgraded: 2026-03-10T10:11:01.432 INFO:teuthology.orchestra.run.vm09.stdout: librados2 librbd1 2026-03-10T10:11:01.485 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:11:01.485 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:11:01.526 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:11:01.527 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:11:01.635 INFO:teuthology.orchestra.run.vm09.stdout:2 upgraded, 107 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:11:01.635 INFO:teuthology.orchestra.run.vm09.stdout:Need to get 178 MB of archives. 2026-03-10T10:11:01.635 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 782 MB of additional disk space will be used. 2026-03-10T10:11:01.635 INFO:teuthology.orchestra.run.vm09.stdout:Get:1 https://archive.ubuntu.com/ubuntu jammy/main amd64 liblttng-ust1 amd64 2.13.1-1ubuntu1 [190 kB] 2026-03-10T10:11:01.673 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:11:01.673 INFO:teuthology.orchestra.run.vm00.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T10:11:01.673 INFO:teuthology.orchestra.run.vm00.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-10T10:11:01.673 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:11:01.674 INFO:teuthology.orchestra.run.vm00.stdout:The following additional packages will be installed: 2026-03-10T10:11:01.674 INFO:teuthology.orchestra.run.vm00.stdout: ceph-base ceph-mgr-cephadm ceph-mgr-dashboard ceph-mgr-diskprediction-local 2026-03-10T10:11:01.674 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-k8sevents ceph-mgr-modules-core ceph-mon ceph-osd jq 2026-03-10T10:11:01.674 INFO:teuthology.orchestra.run.vm00.stdout: libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T10:11:01.674 INFO:teuthology.orchestra.run.vm00.stdout: liboath0 libonig5 libpcre2-16-0 libqt5core5a libqt5dbus5 libqt5network5 2026-03-10T10:11:01.674 INFO:teuthology.orchestra.run.vm00.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsqlite3-mod-ceph 2026-03-10T10:11:01.675 INFO:teuthology.orchestra.run.vm00.stdout: libthrift-0.16.0 lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:11:01.675 INFO:teuthology.orchestra.run.vm00.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:11:01.675 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:11:01.675 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-10T10:11:01.675 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:11:01.675 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:11:01.675 INFO:teuthology.orchestra.run.vm00.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:11:01.675 INFO:teuthology.orchestra.run.vm00.stdout: python3-pastescript python3-pecan python3-pluggy python3-portend 2026-03-10T10:11:01.675 INFO:teuthology.orchestra.run.vm00.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-10T10:11:01.675 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyinotify python3-pytest python3-repoze.lru 2026-03-10T10:11:01.675 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:11:01.675 INFO:teuthology.orchestra.run.vm00.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:11:01.675 INFO:teuthology.orchestra.run.vm00.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:11:01.676 INFO:teuthology.orchestra.run.vm00.stdout: python3-toml python3-waitress python3-wcwidth python3-webob 2026-03-10T10:11:01.676 INFO:teuthology.orchestra.run.vm00.stdout: python3-websocket python3-webtest python3-werkzeug python3-zc.lockfile 2026-03-10T10:11:01.676 INFO:teuthology.orchestra.run.vm00.stdout: qttranslations5-l10n smartmontools socat unzip xmlstarlet zip 2026-03-10T10:11:01.676 INFO:teuthology.orchestra.run.vm00.stdout:Suggested packages: 2026-03-10T10:11:01.676 INFO:teuthology.orchestra.run.vm00.stdout: python3-influxdb readline-doc python3-beaker python-mako-doc 2026-03-10T10:11:01.676 INFO:teuthology.orchestra.run.vm00.stdout: python-natsort-doc httpd-wsgi libapache2-mod-python libapache2-mod-scgi 2026-03-10T10:11:01.676 INFO:teuthology.orchestra.run.vm00.stdout: libjs-mochikit python-pecan-doc python-psutil-doc subversion 2026-03-10T10:11:01.676 INFO:teuthology.orchestra.run.vm00.stdout: python-pygments-doc ttf-bitstream-vera python-pyinotify-doc python3-dap 2026-03-10T10:11:01.676 INFO:teuthology.orchestra.run.vm00.stdout: python-sklearn-doc ipython3 python-waitress-doc python-webob-doc 2026-03-10T10:11:01.676 INFO:teuthology.orchestra.run.vm00.stdout: python-webtest-doc python-werkzeug-doc python3-watchdog gsmartcontrol 2026-03-10T10:11:01.676 INFO:teuthology.orchestra.run.vm00.stdout: smart-notifier mailx | mailutils 2026-03-10T10:11:01.676 INFO:teuthology.orchestra.run.vm00.stdout:Recommended packages: 2026-03-10T10:11:01.676 INFO:teuthology.orchestra.run.vm00.stdout: btrfs-tools 2026-03-10T10:11:01.715 INFO:teuthology.orchestra.run.vm00.stdout:The following NEW packages will be installed: 2026-03-10T10:11:01.715 INFO:teuthology.orchestra.run.vm00.stdout: ceph ceph-base ceph-common ceph-fuse ceph-mds ceph-mgr ceph-mgr-cephadm 2026-03-10T10:11:01.715 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-k8sevents 2026-03-10T10:11:01.715 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core ceph-mon ceph-osd ceph-test ceph-volume cephadm jq 2026-03-10T10:11:01.715 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs-dev libcephfs2 libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 2026-03-10T10:11:01.715 INFO:teuthology.orchestra.run.vm00.stdout: liblua5.3-dev libnbd0 liboath0 libonig5 libpcre2-16-0 libqt5core5a 2026-03-10T10:11:01.715 INFO:teuthology.orchestra.run.vm00.stdout: libqt5dbus5 libqt5network5 libradosstriper1 librdkafka1 libreadline-dev 2026-03-10T10:11:01.715 INFO:teuthology.orchestra.run.vm00.stdout: librgw2 libsqlite3-mod-ceph libthrift-0.16.0 lua-any lua-sec lua-socket 2026-03-10T10:11:01.716 INFO:teuthology.orchestra.run.vm00.stdout: lua5.1 luarocks nvme-cli pkg-config python-asyncssh-doc 2026-03-10T10:11:01.716 INFO:teuthology.orchestra.run.vm00.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:11:01.716 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-argparse python3-ceph-common python3-cephfs python3-cheroot 2026-03-10T10:11:01.716 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-10T10:11:01.716 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:11:01.716 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:11:01.716 INFO:teuthology.orchestra.run.vm00.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:11:01.716 INFO:teuthology.orchestra.run.vm00.stdout: python3-pastescript python3-pecan python3-pluggy python3-portend 2026-03-10T10:11:01.716 INFO:teuthology.orchestra.run.vm00.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-10T10:11:01.716 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyinotify python3-pytest python3-rados python3-rbd 2026-03-10T10:11:01.716 INFO:teuthology.orchestra.run.vm00.stdout: python3-repoze.lru python3-requests-oauthlib python3-rgw python3-routes 2026-03-10T10:11:01.716 INFO:teuthology.orchestra.run.vm00.stdout: python3-rsa python3-simplegeneric python3-simplejson python3-singledispatch 2026-03-10T10:11:01.716 INFO:teuthology.orchestra.run.vm00.stdout: python3-sklearn python3-sklearn-lib python3-tempita python3-tempora 2026-03-10T10:11:01.716 INFO:teuthology.orchestra.run.vm00.stdout: python3-threadpoolctl python3-toml python3-waitress python3-wcwidth 2026-03-10T10:11:01.716 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:11:01.716 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc.lockfile qttranslations5-l10n radosgw rbd-fuse smartmontools 2026-03-10T10:11:01.716 INFO:teuthology.orchestra.run.vm00.stdout: socat unzip xmlstarlet zip 2026-03-10T10:11:01.717 INFO:teuthology.orchestra.run.vm00.stdout:The following packages will be upgraded: 2026-03-10T10:11:01.718 INFO:teuthology.orchestra.run.vm00.stdout: librados2 librbd1 2026-03-10T10:11:01.729 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:11:01.729 INFO:teuthology.orchestra.run.vm03.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T10:11:01.729 INFO:teuthology.orchestra.run.vm03.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-10T10:11:01.729 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:11:01.729 INFO:teuthology.orchestra.run.vm03.stdout:The following additional packages will be installed: 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: ceph-base ceph-mgr-cephadm ceph-mgr-dashboard ceph-mgr-diskprediction-local 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-k8sevents ceph-mgr-modules-core ceph-mon ceph-osd jq 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: liboath0 libonig5 libpcre2-16-0 libqt5core5a libqt5dbus5 libqt5network5 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsqlite3-mod-ceph 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: libthrift-0.16.0 lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python3-pastescript python3-pecan python3-pluggy python3-portend 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyinotify python3-pytest python3-repoze.lru 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python3-toml python3-waitress python3-wcwidth python3-webob 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python3-websocket python3-webtest python3-werkzeug python3-zc.lockfile 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: qttranslations5-l10n smartmontools socat unzip xmlstarlet zip 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout:Suggested packages: 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python3-influxdb readline-doc python3-beaker python-mako-doc 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python-natsort-doc httpd-wsgi libapache2-mod-python libapache2-mod-scgi 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: libjs-mochikit python-pecan-doc python-psutil-doc subversion 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python-pygments-doc ttf-bitstream-vera python-pyinotify-doc python3-dap 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python-sklearn-doc ipython3 python-waitress-doc python-webob-doc 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: python-webtest-doc python-werkzeug-doc python3-watchdog gsmartcontrol 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout: smart-notifier mailx | mailutils 2026-03-10T10:11:01.730 INFO:teuthology.orchestra.run.vm03.stdout:Recommended packages: 2026-03-10T10:11:01.731 INFO:teuthology.orchestra.run.vm03.stdout: btrfs-tools 2026-03-10T10:11:01.767 INFO:teuthology.orchestra.run.vm03.stdout:The following NEW packages will be installed: 2026-03-10T10:11:01.767 INFO:teuthology.orchestra.run.vm03.stdout: ceph ceph-base ceph-common ceph-fuse ceph-mds ceph-mgr ceph-mgr-cephadm 2026-03-10T10:11:01.767 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-k8sevents 2026-03-10T10:11:01.767 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core ceph-mon ceph-osd ceph-test ceph-volume cephadm jq 2026-03-10T10:11:01.767 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs-dev libcephfs2 libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 2026-03-10T10:11:01.767 INFO:teuthology.orchestra.run.vm03.stdout: liblua5.3-dev libnbd0 liboath0 libonig5 libpcre2-16-0 libqt5core5a 2026-03-10T10:11:01.767 INFO:teuthology.orchestra.run.vm03.stdout: libqt5dbus5 libqt5network5 libradosstriper1 librdkafka1 libreadline-dev 2026-03-10T10:11:01.767 INFO:teuthology.orchestra.run.vm03.stdout: librgw2 libsqlite3-mod-ceph libthrift-0.16.0 lua-any lua-sec lua-socket 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: lua5.1 luarocks nvme-cli pkg-config python-asyncssh-doc 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-argparse python3-ceph-common python3-cephfs python3-cheroot 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: python3-pastescript python3-pecan python3-pluggy python3-portend 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyinotify python3-pytest python3-rados python3-rbd 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: python3-repoze.lru python3-requests-oauthlib python3-rgw python3-routes 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: python3-rsa python3-simplegeneric python3-simplejson python3-singledispatch 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: python3-sklearn python3-sklearn-lib python3-tempita python3-tempora 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: python3-threadpoolctl python3-toml python3-waitress python3-wcwidth 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc.lockfile qttranslations5-l10n radosgw rbd-fuse smartmontools 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: socat unzip xmlstarlet zip 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout:The following packages will be upgraded: 2026-03-10T10:11:01.768 INFO:teuthology.orchestra.run.vm03.stdout: librados2 librbd1 2026-03-10T10:11:01.805 INFO:teuthology.orchestra.run.vm09.stdout:Get:2 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libdouble-conversion3 amd64 3.1.7-4 [39.0 kB] 2026-03-10T10:11:01.810 INFO:teuthology.orchestra.run.vm09.stdout:Get:3 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libpcre2-16-0 amd64 10.39-3ubuntu0.1 [203 kB] 2026-03-10T10:11:01.815 INFO:teuthology.orchestra.run.vm00.stdout:2 upgraded, 107 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:11:01.815 INFO:teuthology.orchestra.run.vm00.stdout:Need to get 178 MB of archives. 2026-03-10T10:11:01.815 INFO:teuthology.orchestra.run.vm00.stdout:After this operation, 782 MB of additional disk space will be used. 2026-03-10T10:11:01.815 INFO:teuthology.orchestra.run.vm00.stdout:Get:1 https://archive.ubuntu.com/ubuntu jammy/main amd64 liblttng-ust1 amd64 2.13.1-1ubuntu1 [190 kB] 2026-03-10T10:11:01.845 INFO:teuthology.orchestra.run.vm09.stdout:Get:4 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5core5a amd64 5.15.3+dfsg-2ubuntu0.2 [2006 kB] 2026-03-10T10:11:01.852 INFO:teuthology.orchestra.run.vm00.stdout:Get:2 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libdouble-conversion3 amd64 3.1.7-4 [39.0 kB] 2026-03-10T10:11:01.853 INFO:teuthology.orchestra.run.vm00.stdout:Get:3 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libpcre2-16-0 amd64 10.39-3ubuntu0.1 [203 kB] 2026-03-10T10:11:01.860 INFO:teuthology.orchestra.run.vm00.stdout:Get:4 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5core5a amd64 5.15.3+dfsg-2ubuntu0.2 [2006 kB] 2026-03-10T10:11:01.887 INFO:teuthology.orchestra.run.vm00.stdout:Get:5 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5dbus5 amd64 5.15.3+dfsg-2ubuntu0.2 [222 kB] 2026-03-10T10:11:01.889 INFO:teuthology.orchestra.run.vm00.stdout:Get:6 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5network5 amd64 5.15.3+dfsg-2ubuntu0.2 [731 kB] 2026-03-10T10:11:01.903 INFO:teuthology.orchestra.run.vm00.stdout:Get:7 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libthrift-0.16.0 amd64 0.16.0-2 [267 kB] 2026-03-10T10:11:01.905 INFO:teuthology.orchestra.run.vm00.stdout:Get:8 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libnbd0 amd64 1.10.5-1 [71.3 kB] 2026-03-10T10:11:01.905 INFO:teuthology.orchestra.run.vm00.stdout:Get:9 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-wcwidth all 0.2.5+dfsg1-1 [21.9 kB] 2026-03-10T10:11:01.905 INFO:teuthology.orchestra.run.vm00.stdout:Get:10 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-prettytable all 2.5.0-2 [31.3 kB] 2026-03-10T10:11:01.906 INFO:teuthology.orchestra.run.vm00.stdout:Get:11 https://archive.ubuntu.com/ubuntu jammy/universe amd64 librdkafka1 amd64 1.8.0-1build1 [633 kB] 2026-03-10T10:11:01.909 INFO:teuthology.orchestra.run.vm00.stdout:Get:12 https://archive.ubuntu.com/ubuntu jammy/main amd64 libreadline-dev amd64 8.1.2-1 [166 kB] 2026-03-10T10:11:01.910 INFO:teuthology.orchestra.run.vm00.stdout:Get:13 https://archive.ubuntu.com/ubuntu jammy/main amd64 liblua5.3-dev amd64 5.3.6-1build1 [167 kB] 2026-03-10T10:11:01.911 INFO:teuthology.orchestra.run.vm00.stdout:Get:14 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua5.1 amd64 5.1.5-8.1build4 [94.6 kB] 2026-03-10T10:11:01.911 INFO:teuthology.orchestra.run.vm00.stdout:Get:15 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-any all 27ubuntu1 [5034 B] 2026-03-10T10:11:01.916 INFO:teuthology.orchestra.run.vm00.stdout:Get:16 https://archive.ubuntu.com/ubuntu jammy/main amd64 zip amd64 3.0-12build2 [176 kB] 2026-03-10T10:11:01.918 INFO:teuthology.orchestra.run.vm00.stdout:Get:17 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 unzip amd64 6.0-26ubuntu3.2 [175 kB] 2026-03-10T10:11:01.920 INFO:teuthology.orchestra.run.vm00.stdout:Get:18 https://archive.ubuntu.com/ubuntu jammy/universe amd64 luarocks all 3.8.0+dfsg1-1 [140 kB] 2026-03-10T10:11:01.921 INFO:teuthology.orchestra.run.vm00.stdout:Get:19 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 liboath0 amd64 2.6.7-3ubuntu0.1 [41.3 kB] 2026-03-10T10:11:01.921 INFO:teuthology.orchestra.run.vm00.stdout:Get:20 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.functools all 3.4.0-2 [9030 B] 2026-03-10T10:11:01.925 INFO:teuthology.orchestra.run.vm00.stdout:Get:21 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-cheroot all 8.5.2+ds1-1ubuntu3.1 [71.1 kB] 2026-03-10T10:11:01.925 INFO:teuthology.orchestra.run.vm00.stdout:Get:22 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.classes all 3.2.1-3 [6452 B] 2026-03-10T10:11:01.926 INFO:teuthology.orchestra.run.vm00.stdout:Get:23 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.text all 3.6.0-2 [8716 B] 2026-03-10T10:11:01.926 INFO:teuthology.orchestra.run.vm00.stdout:Get:24 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.collections all 3.4.0-2 [11.4 kB] 2026-03-10T10:11:01.926 INFO:teuthology.orchestra.run.vm00.stdout:Get:25 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempora all 4.1.2-1 [14.8 kB] 2026-03-10T10:11:01.933 INFO:teuthology.orchestra.run.vm00.stdout:Get:26 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-portend all 3.0.0-1 [7240 B] 2026-03-10T10:11:01.933 INFO:teuthology.orchestra.run.vm00.stdout:Get:27 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-zc.lockfile all 2.0-1 [8980 B] 2026-03-10T10:11:01.933 INFO:teuthology.orchestra.run.vm00.stdout:Get:28 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cherrypy3 all 18.6.1-4 [208 kB] 2026-03-10T10:11:01.935 INFO:teuthology.orchestra.run.vm00.stdout:Get:29 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-natsort all 8.0.2-1 [35.3 kB] 2026-03-10T10:11:01.935 INFO:teuthology.orchestra.run.vm00.stdout:Get:30 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-logutils all 0.3.3-8 [17.6 kB] 2026-03-10T10:11:01.941 INFO:teuthology.orchestra.run.vm00.stdout:Get:31 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-mako all 1.1.3+ds1-2ubuntu0.1 [60.5 kB] 2026-03-10T10:11:01.941 INFO:teuthology.orchestra.run.vm00.stdout:Get:32 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplegeneric all 0.8.1-3 [11.3 kB] 2026-03-10T10:11:01.941 INFO:teuthology.orchestra.run.vm00.stdout:Get:33 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-singledispatch all 3.4.0.3-3 [7320 B] 2026-03-10T10:11:01.942 INFO:teuthology.orchestra.run.vm00.stdout:Get:34 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-webob all 1:1.8.6-1.1ubuntu0.1 [86.7 kB] 2026-03-10T10:11:01.942 INFO:teuthology.orchestra.run.vm00.stdout:Get:35 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-waitress all 1.4.4-1.1ubuntu1.1 [47.0 kB] 2026-03-10T10:11:01.948 INFO:teuthology.orchestra.run.vm00.stdout:Get:36 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempita all 0.5.2-6ubuntu1 [15.1 kB] 2026-03-10T10:11:01.948 INFO:teuthology.orchestra.run.vm00.stdout:Get:37 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-paste all 3.5.0+dfsg1-1 [456 kB] 2026-03-10T10:11:01.952 INFO:teuthology.orchestra.run.vm00.stdout:Get:38 https://archive.ubuntu.com/ubuntu jammy/main amd64 python-pastedeploy-tpl all 2.1.1-1 [4892 B] 2026-03-10T10:11:01.952 INFO:teuthology.orchestra.run.vm00.stdout:Get:39 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pastedeploy all 2.1.1-1 [26.6 kB] 2026-03-10T10:11:01.953 INFO:teuthology.orchestra.run.vm00.stdout:Get:40 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-webtest all 2.0.35-1 [28.5 kB] 2026-03-10T10:11:01.956 INFO:teuthology.orchestra.run.vm00.stdout:Get:41 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pecan all 1.3.3-4ubuntu2 [87.3 kB] 2026-03-10T10:11:01.957 INFO:teuthology.orchestra.run.vm00.stdout:Get:42 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-werkzeug all 2.0.2+dfsg1-1ubuntu0.22.04.3 [181 kB] 2026-03-10T10:11:01.958 INFO:teuthology.orchestra.run.vm00.stdout:Get:43 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libfuse2 amd64 2.9.9-5ubuntu3 [90.3 kB] 2026-03-10T10:11:01.959 INFO:teuthology.orchestra.run.vm00.stdout:Get:44 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3-asyncssh all 2.5.0-1ubuntu0.1 [189 kB] 2026-03-10T10:11:01.961 INFO:teuthology.orchestra.run.vm00.stdout:Get:45 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-repoze.lru all 0.7-2 [12.1 kB] 2026-03-10T10:11:01.963 INFO:teuthology.orchestra.run.vm09.stdout:Get:5 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5dbus5 amd64 5.15.3+dfsg-2ubuntu0.2 [222 kB] 2026-03-10T10:11:01.964 INFO:teuthology.orchestra.run.vm00.stdout:Get:46 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-routes all 2.5.1-1ubuntu1 [89.0 kB] 2026-03-10T10:11:01.965 INFO:teuthology.orchestra.run.vm00.stdout:Get:47 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn-lib amd64 0.23.2-5ubuntu6 [2058 kB] 2026-03-10T10:11:02.010 INFO:teuthology.orchestra.run.vm00.stdout:Get:48 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-joblib all 0.17.0-4ubuntu1 [204 kB] 2026-03-10T10:11:02.010 INFO:teuthology.orchestra.run.vm00.stdout:Get:49 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-threadpoolctl all 3.1.0-1 [21.3 kB] 2026-03-10T10:11:02.011 INFO:teuthology.orchestra.run.vm00.stdout:Get:50 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn all 0.23.2-5ubuntu6 [1829 kB] 2026-03-10T10:11:02.015 INFO:teuthology.orchestra.run.vm09.stdout:Get:6 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5network5 amd64 5.15.3+dfsg-2ubuntu0.2 [731 kB] 2026-03-10T10:11:02.018 INFO:teuthology.orchestra.run.vm00.stdout:Get:51 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cachetools all 5.0.0-1 [9722 B] 2026-03-10T10:11:02.018 INFO:teuthology.orchestra.run.vm00.stdout:Get:52 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-rsa all 4.8-1 [28.4 kB] 2026-03-10T10:11:02.018 INFO:teuthology.orchestra.run.vm00.stdout:Get:53 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-google-auth all 1.5.1-3 [35.7 kB] 2026-03-10T10:11:02.018 INFO:teuthology.orchestra.run.vm00.stdout:Get:54 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-requests-oauthlib all 1.3.0+ds-0.1 [18.7 kB] 2026-03-10T10:11:02.019 INFO:teuthology.orchestra.run.vm00.stdout:Get:55 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-websocket all 1.2.3-1 [34.7 kB] 2026-03-10T10:11:02.019 INFO:teuthology.orchestra.run.vm00.stdout:Get:56 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-kubernetes all 12.0.1-1ubuntu1 [353 kB] 2026-03-10T10:11:02.021 INFO:teuthology.orchestra.run.vm00.stdout:Get:57 https://archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB] 2026-03-10T10:11:02.028 INFO:teuthology.orchestra.run.vm00.stdout:Get:58 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB] 2026-03-10T10:11:02.029 INFO:teuthology.orchestra.run.vm00.stdout:Get:59 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB] 2026-03-10T10:11:02.029 INFO:teuthology.orchestra.run.vm00.stdout:Get:60 https://archive.ubuntu.com/ubuntu jammy/main amd64 socat amd64 1.7.4.1-3ubuntu4 [349 kB] 2026-03-10T10:11:02.033 INFO:teuthology.orchestra.run.vm00.stdout:Get:61 https://archive.ubuntu.com/ubuntu jammy/universe amd64 xmlstarlet amd64 1.6.1-2.1 [265 kB] 2026-03-10T10:11:02.033 INFO:teuthology.orchestra.run.vm09.stdout:Get:7 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libthrift-0.16.0 amd64 0.16.0-2 [267 kB] 2026-03-10T10:11:02.035 INFO:teuthology.orchestra.run.vm00.stdout:Get:62 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-socket amd64 3.0~rc1+git+ac3201d-6 [78.9 kB] 2026-03-10T10:11:02.036 INFO:teuthology.orchestra.run.vm00.stdout:Get:63 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-sec amd64 1.0.2-1 [37.6 kB] 2026-03-10T10:11:02.036 INFO:teuthology.orchestra.run.vm00.stdout:Get:64 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 nvme-cli amd64 1.16-3ubuntu0.3 [474 kB] 2026-03-10T10:11:02.041 INFO:teuthology.orchestra.run.vm09.stdout:Get:8 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libnbd0 amd64 1.10.5-1 [71.3 kB] 2026-03-10T10:11:02.042 INFO:teuthology.orchestra.run.vm09.stdout:Get:9 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-wcwidth all 0.2.5+dfsg1-1 [21.9 kB] 2026-03-10T10:11:02.042 INFO:teuthology.orchestra.run.vm09.stdout:Get:10 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-prettytable all 2.5.0-2 [31.3 kB] 2026-03-10T10:11:02.043 INFO:teuthology.orchestra.run.vm09.stdout:Get:11 https://archive.ubuntu.com/ubuntu jammy/universe amd64 librdkafka1 amd64 1.8.0-1build1 [633 kB] 2026-03-10T10:11:02.046 INFO:teuthology.orchestra.run.vm09.stdout:Get:12 https://archive.ubuntu.com/ubuntu jammy/main amd64 libreadline-dev amd64 8.1.2-1 [166 kB] 2026-03-10T10:11:02.055 INFO:teuthology.orchestra.run.vm00.stdout:Get:65 https://archive.ubuntu.com/ubuntu jammy/main amd64 pkg-config amd64 0.29.2-1ubuntu3 [48.2 kB] 2026-03-10T10:11:02.055 INFO:teuthology.orchestra.run.vm00.stdout:Get:66 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python-asyncssh-doc all 2.5.0-1ubuntu0.1 [309 kB] 2026-03-10T10:11:02.059 INFO:teuthology.orchestra.run.vm00.stdout:Get:67 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-iniconfig all 1.1.1-2 [6024 B] 2026-03-10T10:11:02.059 INFO:teuthology.orchestra.run.vm00.stdout:Get:68 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pastescript all 2.0.2-4 [54.6 kB] 2026-03-10T10:11:02.059 INFO:teuthology.orchestra.run.vm00.stdout:Get:69 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pluggy all 0.13.0-7.1 [19.0 kB] 2026-03-10T10:11:02.059 INFO:teuthology.orchestra.run.vm00.stdout:Get:70 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-psutil amd64 5.9.0-1build1 [158 kB] 2026-03-10T10:11:02.063 INFO:teuthology.orchestra.run.vm00.stdout:Get:71 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-py all 1.10.0-1 [71.9 kB] 2026-03-10T10:11:02.063 INFO:teuthology.orchestra.run.vm00.stdout:Get:72 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-pygments all 2.11.2+dfsg-2ubuntu0.1 [750 kB] 2026-03-10T10:11:02.070 INFO:teuthology.orchestra.run.vm00.stdout:Get:73 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pyinotify all 0.9.6-1.3 [24.8 kB] 2026-03-10T10:11:02.070 INFO:teuthology.orchestra.run.vm00.stdout:Get:74 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-toml all 0.10.2-1 [16.5 kB] 2026-03-10T10:11:02.072 INFO:teuthology.orchestra.run.vm00.stdout:Get:75 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pytest all 6.2.5-1ubuntu2 [214 kB] 2026-03-10T10:11:02.074 INFO:teuthology.orchestra.run.vm00.stdout:Get:76 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplejson amd64 3.17.6-1build1 [54.7 kB] 2026-03-10T10:11:02.075 INFO:teuthology.orchestra.run.vm00.stdout:Get:77 https://archive.ubuntu.com/ubuntu jammy/universe amd64 qttranslations5-l10n all 5.15.3-1 [1983 kB] 2026-03-10T10:11:02.102 INFO:teuthology.orchestra.run.vm00.stdout:Get:78 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 smartmontools amd64 7.2-1ubuntu0.1 [583 kB] 2026-03-10T10:11:02.124 INFO:teuthology.orchestra.run.vm09.stdout:Get:13 https://archive.ubuntu.com/ubuntu jammy/main amd64 liblua5.3-dev amd64 5.3.6-1build1 [167 kB] 2026-03-10T10:11:02.125 INFO:teuthology.orchestra.run.vm09.stdout:Get:14 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua5.1 amd64 5.1.5-8.1build4 [94.6 kB] 2026-03-10T10:11:02.126 INFO:teuthology.orchestra.run.vm09.stdout:Get:15 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-any all 27ubuntu1 [5034 B] 2026-03-10T10:11:02.126 INFO:teuthology.orchestra.run.vm09.stdout:Get:16 https://archive.ubuntu.com/ubuntu jammy/main amd64 zip amd64 3.0-12build2 [176 kB] 2026-03-10T10:11:02.127 INFO:teuthology.orchestra.run.vm09.stdout:Get:17 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 unzip amd64 6.0-26ubuntu3.2 [175 kB] 2026-03-10T10:11:02.128 INFO:teuthology.orchestra.run.vm09.stdout:Get:18 https://archive.ubuntu.com/ubuntu jammy/universe amd64 luarocks all 3.8.0+dfsg1-1 [140 kB] 2026-03-10T10:11:02.129 INFO:teuthology.orchestra.run.vm09.stdout:Get:19 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 liboath0 amd64 2.6.7-3ubuntu0.1 [41.3 kB] 2026-03-10T10:11:02.129 INFO:teuthology.orchestra.run.vm09.stdout:Get:20 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.functools all 3.4.0-2 [9030 B] 2026-03-10T10:11:02.130 INFO:teuthology.orchestra.run.vm09.stdout:Get:21 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-cheroot all 8.5.2+ds1-1ubuntu3.1 [71.1 kB] 2026-03-10T10:11:02.130 INFO:teuthology.orchestra.run.vm09.stdout:Get:22 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.classes all 3.2.1-3 [6452 B] 2026-03-10T10:11:02.162 INFO:teuthology.orchestra.run.vm09.stdout:Get:23 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.text all 3.6.0-2 [8716 B] 2026-03-10T10:11:02.166 INFO:teuthology.orchestra.run.vm09.stdout:Get:24 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.collections all 3.4.0-2 [11.4 kB] 2026-03-10T10:11:02.166 INFO:teuthology.orchestra.run.vm09.stdout:Get:25 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempora all 4.1.2-1 [14.8 kB] 2026-03-10T10:11:02.167 INFO:teuthology.orchestra.run.vm09.stdout:Get:26 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-portend all 3.0.0-1 [7240 B] 2026-03-10T10:11:02.167 INFO:teuthology.orchestra.run.vm09.stdout:Get:27 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-zc.lockfile all 2.0-1 [8980 B] 2026-03-10T10:11:02.168 INFO:teuthology.orchestra.run.vm09.stdout:Get:28 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cherrypy3 all 18.6.1-4 [208 kB] 2026-03-10T10:11:02.169 INFO:teuthology.orchestra.run.vm09.stdout:Get:29 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-natsort all 8.0.2-1 [35.3 kB] 2026-03-10T10:11:02.170 INFO:teuthology.orchestra.run.vm09.stdout:Get:30 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-logutils all 0.3.3-8 [17.6 kB] 2026-03-10T10:11:02.198 INFO:teuthology.orchestra.run.vm09.stdout:Get:31 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-mako all 1.1.3+ds1-2ubuntu0.1 [60.5 kB] 2026-03-10T10:11:02.198 INFO:teuthology.orchestra.run.vm09.stdout:Get:32 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplegeneric all 0.8.1-3 [11.3 kB] 2026-03-10T10:11:02.206 INFO:teuthology.orchestra.run.vm09.stdout:Get:33 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-singledispatch all 3.4.0.3-3 [7320 B] 2026-03-10T10:11:02.206 INFO:teuthology.orchestra.run.vm09.stdout:Get:34 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-webob all 1:1.8.6-1.1ubuntu0.1 [86.7 kB] 2026-03-10T10:11:02.207 INFO:teuthology.orchestra.run.vm09.stdout:Get:35 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-waitress all 1.4.4-1.1ubuntu1.1 [47.0 kB] 2026-03-10T10:11:02.207 INFO:teuthology.orchestra.run.vm09.stdout:Get:36 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempita all 0.5.2-6ubuntu1 [15.1 kB] 2026-03-10T10:11:02.208 INFO:teuthology.orchestra.run.vm09.stdout:Get:37 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-paste all 3.5.0+dfsg1-1 [456 kB] 2026-03-10T10:11:02.213 INFO:teuthology.orchestra.run.vm09.stdout:Get:38 https://archive.ubuntu.com/ubuntu jammy/main amd64 python-pastedeploy-tpl all 2.1.1-1 [4892 B] 2026-03-10T10:11:02.213 INFO:teuthology.orchestra.run.vm09.stdout:Get:39 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pastedeploy all 2.1.1-1 [26.6 kB] 2026-03-10T10:11:02.216 INFO:teuthology.orchestra.run.vm09.stdout:Get:40 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librbd1 amd64 19.2.3-678-ge911bdeb-1jammy [3257 kB] 2026-03-10T10:11:02.233 INFO:teuthology.orchestra.run.vm09.stdout:Get:41 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-webtest all 2.0.35-1 [28.5 kB] 2026-03-10T10:11:02.238 INFO:teuthology.orchestra.run.vm03.stdout:2 upgraded, 107 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:11:02.238 INFO:teuthology.orchestra.run.vm03.stdout:Need to get 178 MB of archives. 2026-03-10T10:11:02.238 INFO:teuthology.orchestra.run.vm03.stdout:After this operation, 782 MB of additional disk space will be used. 2026-03-10T10:11:02.238 INFO:teuthology.orchestra.run.vm03.stdout:Get:1 https://archive.ubuntu.com/ubuntu jammy/main amd64 liblttng-ust1 amd64 2.13.1-1ubuntu1 [190 kB] 2026-03-10T10:11:02.249 INFO:teuthology.orchestra.run.vm09.stdout:Get:42 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pecan all 1.3.3-4ubuntu2 [87.3 kB] 2026-03-10T10:11:02.250 INFO:teuthology.orchestra.run.vm09.stdout:Get:43 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-werkzeug all 2.0.2+dfsg1-1ubuntu0.22.04.3 [181 kB] 2026-03-10T10:11:02.252 INFO:teuthology.orchestra.run.vm09.stdout:Get:44 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libfuse2 amd64 2.9.9-5ubuntu3 [90.3 kB] 2026-03-10T10:11:02.254 INFO:teuthology.orchestra.run.vm09.stdout:Get:45 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3-asyncssh all 2.5.0-1ubuntu0.1 [189 kB] 2026-03-10T10:11:02.256 INFO:teuthology.orchestra.run.vm09.stdout:Get:46 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-repoze.lru all 0.7-2 [12.1 kB] 2026-03-10T10:11:02.256 INFO:teuthology.orchestra.run.vm09.stdout:Get:47 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-routes all 2.5.1-1ubuntu1 [89.0 kB] 2026-03-10T10:11:02.257 INFO:teuthology.orchestra.run.vm09.stdout:Get:48 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn-lib amd64 0.23.2-5ubuntu6 [2058 kB] 2026-03-10T10:11:02.281 INFO:teuthology.orchestra.run.vm09.stdout:Get:49 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-joblib all 0.17.0-4ubuntu1 [204 kB] 2026-03-10T10:11:02.283 INFO:teuthology.orchestra.run.vm09.stdout:Get:50 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-threadpoolctl all 3.1.0-1 [21.3 kB] 2026-03-10T10:11:02.306 INFO:teuthology.orchestra.run.vm09.stdout:Get:51 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn all 0.23.2-5ubuntu6 [1829 kB] 2026-03-10T10:11:02.322 INFO:teuthology.orchestra.run.vm09.stdout:Get:52 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cachetools all 5.0.0-1 [9722 B] 2026-03-10T10:11:02.322 INFO:teuthology.orchestra.run.vm09.stdout:Get:53 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-rsa all 4.8-1 [28.4 kB] 2026-03-10T10:11:02.323 INFO:teuthology.orchestra.run.vm09.stdout:Get:54 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-google-auth all 1.5.1-3 [35.7 kB] 2026-03-10T10:11:02.323 INFO:teuthology.orchestra.run.vm09.stdout:Get:55 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-requests-oauthlib all 1.3.0+ds-0.1 [18.7 kB] 2026-03-10T10:11:02.324 INFO:teuthology.orchestra.run.vm09.stdout:Get:56 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-websocket all 1.2.3-1 [34.7 kB] 2026-03-10T10:11:02.324 INFO:teuthology.orchestra.run.vm09.stdout:Get:57 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-kubernetes all 12.0.1-1ubuntu1 [353 kB] 2026-03-10T10:11:02.342 INFO:teuthology.orchestra.run.vm09.stdout:Get:58 https://archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB] 2026-03-10T10:11:02.344 INFO:teuthology.orchestra.run.vm09.stdout:Get:59 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB] 2026-03-10T10:11:02.345 INFO:teuthology.orchestra.run.vm09.stdout:Get:60 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB] 2026-03-10T10:11:02.378 INFO:teuthology.orchestra.run.vm09.stdout:Get:61 https://archive.ubuntu.com/ubuntu jammy/main amd64 socat amd64 1.7.4.1-3ubuntu4 [349 kB] 2026-03-10T10:11:02.381 INFO:teuthology.orchestra.run.vm09.stdout:Get:62 https://archive.ubuntu.com/ubuntu jammy/universe amd64 xmlstarlet amd64 1.6.1-2.1 [265 kB] 2026-03-10T10:11:02.383 INFO:teuthology.orchestra.run.vm09.stdout:Get:63 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-socket amd64 3.0~rc1+git+ac3201d-6 [78.9 kB] 2026-03-10T10:11:02.384 INFO:teuthology.orchestra.run.vm09.stdout:Get:64 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-sec amd64 1.0.2-1 [37.6 kB] 2026-03-10T10:11:02.384 INFO:teuthology.orchestra.run.vm09.stdout:Get:65 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 nvme-cli amd64 1.16-3ubuntu0.3 [474 kB] 2026-03-10T10:11:02.388 INFO:teuthology.orchestra.run.vm09.stdout:Get:66 https://archive.ubuntu.com/ubuntu jammy/main amd64 pkg-config amd64 0.29.2-1ubuntu3 [48.2 kB] 2026-03-10T10:11:02.389 INFO:teuthology.orchestra.run.vm09.stdout:Get:67 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python-asyncssh-doc all 2.5.0-1ubuntu0.1 [309 kB] 2026-03-10T10:11:02.414 INFO:teuthology.orchestra.run.vm09.stdout:Get:68 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-iniconfig all 1.1.1-2 [6024 B] 2026-03-10T10:11:02.414 INFO:teuthology.orchestra.run.vm09.stdout:Get:69 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pastescript all 2.0.2-4 [54.6 kB] 2026-03-10T10:11:02.415 INFO:teuthology.orchestra.run.vm09.stdout:Get:70 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pluggy all 0.13.0-7.1 [19.0 kB] 2026-03-10T10:11:02.449 INFO:teuthology.orchestra.run.vm09.stdout:Get:71 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-psutil amd64 5.9.0-1build1 [158 kB] 2026-03-10T10:11:02.451 INFO:teuthology.orchestra.run.vm09.stdout:Get:72 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-py all 1.10.0-1 [71.9 kB] 2026-03-10T10:11:02.451 INFO:teuthology.orchestra.run.vm09.stdout:Get:73 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-pygments all 2.11.2+dfsg-2ubuntu0.1 [750 kB] 2026-03-10T10:11:02.458 INFO:teuthology.orchestra.run.vm09.stdout:Get:74 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pyinotify all 0.9.6-1.3 [24.8 kB] 2026-03-10T10:11:02.458 INFO:teuthology.orchestra.run.vm09.stdout:Get:75 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-toml all 0.10.2-1 [16.5 kB] 2026-03-10T10:11:02.458 INFO:teuthology.orchestra.run.vm09.stdout:Get:76 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pytest all 6.2.5-1ubuntu2 [214 kB] 2026-03-10T10:11:02.460 INFO:teuthology.orchestra.run.vm09.stdout:Get:77 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplejson amd64 3.17.6-1build1 [54.7 kB] 2026-03-10T10:11:02.485 INFO:teuthology.orchestra.run.vm09.stdout:Get:78 https://archive.ubuntu.com/ubuntu jammy/universe amd64 qttranslations5-l10n all 5.15.3-1 [1983 kB] 2026-03-10T10:11:02.505 INFO:teuthology.orchestra.run.vm00.stdout:Get:79 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librbd1 amd64 19.2.3-678-ge911bdeb-1jammy [3257 kB] 2026-03-10T10:11:02.554 INFO:teuthology.orchestra.run.vm09.stdout:Get:79 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 smartmontools amd64 7.2-1ubuntu0.1 [583 kB] 2026-03-10T10:11:02.562 INFO:teuthology.orchestra.run.vm03.stdout:Get:2 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librbd1 amd64 19.2.3-678-ge911bdeb-1jammy [3257 kB] 2026-03-10T10:11:02.700 INFO:teuthology.orchestra.run.vm03.stdout:Get:3 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libdouble-conversion3 amd64 3.1.7-4 [39.0 kB] 2026-03-10T10:11:02.715 INFO:teuthology.orchestra.run.vm03.stdout:Get:4 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libpcre2-16-0 amd64 10.39-3ubuntu0.1 [203 kB] 2026-03-10T10:11:02.810 INFO:teuthology.orchestra.run.vm03.stdout:Get:5 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5core5a amd64 5.15.3+dfsg-2ubuntu0.2 [2006 kB] 2026-03-10T10:11:03.085 INFO:teuthology.orchestra.run.vm03.stdout:Get:6 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5dbus5 amd64 5.15.3+dfsg-2ubuntu0.2 [222 kB] 2026-03-10T10:11:03.099 INFO:teuthology.orchestra.run.vm03.stdout:Get:7 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5network5 amd64 5.15.3+dfsg-2ubuntu0.2 [731 kB] 2026-03-10T10:11:03.136 INFO:teuthology.orchestra.run.vm03.stdout:Get:8 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libthrift-0.16.0 amd64 0.16.0-2 [267 kB] 2026-03-10T10:11:03.147 INFO:teuthology.orchestra.run.vm03.stdout:Get:9 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libnbd0 amd64 1.10.5-1 [71.3 kB] 2026-03-10T10:11:03.150 INFO:teuthology.orchestra.run.vm03.stdout:Get:10 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-wcwidth all 0.2.5+dfsg1-1 [21.9 kB] 2026-03-10T10:11:03.150 INFO:teuthology.orchestra.run.vm03.stdout:Get:11 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-prettytable all 2.5.0-2 [31.3 kB] 2026-03-10T10:11:03.152 INFO:teuthology.orchestra.run.vm03.stdout:Get:12 https://archive.ubuntu.com/ubuntu jammy/universe amd64 librdkafka1 amd64 1.8.0-1build1 [633 kB] 2026-03-10T10:11:03.174 INFO:teuthology.orchestra.run.vm03.stdout:Get:13 https://archive.ubuntu.com/ubuntu jammy/main amd64 libreadline-dev amd64 8.1.2-1 [166 kB] 2026-03-10T10:11:03.180 INFO:teuthology.orchestra.run.vm03.stdout:Get:14 https://archive.ubuntu.com/ubuntu jammy/main amd64 liblua5.3-dev amd64 5.3.6-1build1 [167 kB] 2026-03-10T10:11:03.185 INFO:teuthology.orchestra.run.vm03.stdout:Get:15 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua5.1 amd64 5.1.5-8.1build4 [94.6 kB] 2026-03-10T10:11:03.248 INFO:teuthology.orchestra.run.vm09.stdout:Get:80 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librados2 amd64 19.2.3-678-ge911bdeb-1jammy [3597 kB] 2026-03-10T10:11:03.279 INFO:teuthology.orchestra.run.vm03.stdout:Get:16 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-any all 27ubuntu1 [5034 B] 2026-03-10T10:11:03.279 INFO:teuthology.orchestra.run.vm03.stdout:Get:17 https://archive.ubuntu.com/ubuntu jammy/main amd64 zip amd64 3.0-12build2 [176 kB] 2026-03-10T10:11:03.281 INFO:teuthology.orchestra.run.vm03.stdout:Get:18 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 unzip amd64 6.0-26ubuntu3.2 [175 kB] 2026-03-10T10:11:03.284 INFO:teuthology.orchestra.run.vm03.stdout:Get:19 https://archive.ubuntu.com/ubuntu jammy/universe amd64 luarocks all 3.8.0+dfsg1-1 [140 kB] 2026-03-10T10:11:03.286 INFO:teuthology.orchestra.run.vm03.stdout:Get:20 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 liboath0 amd64 2.6.7-3ubuntu0.1 [41.3 kB] 2026-03-10T10:11:03.287 INFO:teuthology.orchestra.run.vm03.stdout:Get:21 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.functools all 3.4.0-2 [9030 B] 2026-03-10T10:11:03.287 INFO:teuthology.orchestra.run.vm03.stdout:Get:22 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-cheroot all 8.5.2+ds1-1ubuntu3.1 [71.1 kB] 2026-03-10T10:11:03.288 INFO:teuthology.orchestra.run.vm03.stdout:Get:23 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.classes all 3.2.1-3 [6452 B] 2026-03-10T10:11:03.288 INFO:teuthology.orchestra.run.vm03.stdout:Get:24 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.text all 3.6.0-2 [8716 B] 2026-03-10T10:11:03.385 INFO:teuthology.orchestra.run.vm03.stdout:Get:25 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.collections all 3.4.0-2 [11.4 kB] 2026-03-10T10:11:03.385 INFO:teuthology.orchestra.run.vm03.stdout:Get:26 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempora all 4.1.2-1 [14.8 kB] 2026-03-10T10:11:03.385 INFO:teuthology.orchestra.run.vm03.stdout:Get:27 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-portend all 3.0.0-1 [7240 B] 2026-03-10T10:11:03.386 INFO:teuthology.orchestra.run.vm03.stdout:Get:28 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-zc.lockfile all 2.0-1 [8980 B] 2026-03-10T10:11:03.484 INFO:teuthology.orchestra.run.vm03.stdout:Get:29 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cherrypy3 all 18.6.1-4 [208 kB] 2026-03-10T10:11:03.487 INFO:teuthology.orchestra.run.vm03.stdout:Get:30 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-natsort all 8.0.2-1 [35.3 kB] 2026-03-10T10:11:03.487 INFO:teuthology.orchestra.run.vm03.stdout:Get:31 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-logutils all 0.3.3-8 [17.6 kB] 2026-03-10T10:11:03.487 INFO:teuthology.orchestra.run.vm03.stdout:Get:32 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-mako all 1.1.3+ds1-2ubuntu0.1 [60.5 kB] 2026-03-10T10:11:03.488 INFO:teuthology.orchestra.run.vm03.stdout:Get:33 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplegeneric all 0.8.1-3 [11.3 kB] 2026-03-10T10:11:03.488 INFO:teuthology.orchestra.run.vm03.stdout:Get:34 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-singledispatch all 3.4.0.3-3 [7320 B] 2026-03-10T10:11:03.541 INFO:teuthology.orchestra.run.vm09.stdout:Get:81 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs2 amd64 19.2.3-678-ge911bdeb-1jammy [979 kB] 2026-03-10T10:11:03.544 INFO:teuthology.orchestra.run.vm09.stdout:Get:82 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rados amd64 19.2.3-678-ge911bdeb-1jammy [357 kB] 2026-03-10T10:11:03.545 INFO:teuthology.orchestra.run.vm09.stdout:Get:83 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-argparse all 19.2.3-678-ge911bdeb-1jammy [32.9 kB] 2026-03-10T10:11:03.545 INFO:teuthology.orchestra.run.vm09.stdout:Get:84 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-cephfs amd64 19.2.3-678-ge911bdeb-1jammy [184 kB] 2026-03-10T10:11:03.546 INFO:teuthology.orchestra.run.vm09.stdout:Get:85 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-common all 19.2.3-678-ge911bdeb-1jammy [70.1 kB] 2026-03-10T10:11:03.583 INFO:teuthology.orchestra.run.vm03.stdout:Get:35 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-webob all 1:1.8.6-1.1ubuntu0.1 [86.7 kB] 2026-03-10T10:11:03.584 INFO:teuthology.orchestra.run.vm03.stdout:Get:36 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-waitress all 1.4.4-1.1ubuntu1.1 [47.0 kB] 2026-03-10T10:11:03.585 INFO:teuthology.orchestra.run.vm03.stdout:Get:37 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempita all 0.5.2-6ubuntu1 [15.1 kB] 2026-03-10T10:11:03.585 INFO:teuthology.orchestra.run.vm03.stdout:Get:38 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-paste all 3.5.0+dfsg1-1 [456 kB] 2026-03-10T10:11:03.625 INFO:teuthology.orchestra.run.vm00.stdout:Get:80 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librados2 amd64 19.2.3-678-ge911bdeb-1jammy [3597 kB] 2026-03-10T10:11:03.669 INFO:teuthology.orchestra.run.vm09.stdout:Get:86 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rbd amd64 19.2.3-678-ge911bdeb-1jammy [334 kB] 2026-03-10T10:11:03.674 INFO:teuthology.orchestra.run.vm09.stdout:Get:87 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librgw2 amd64 19.2.3-678-ge911bdeb-1jammy [6935 kB] 2026-03-10T10:11:03.682 INFO:teuthology.orchestra.run.vm03.stdout:Get:39 https://archive.ubuntu.com/ubuntu jammy/main amd64 python-pastedeploy-tpl all 2.1.1-1 [4892 B] 2026-03-10T10:11:03.682 INFO:teuthology.orchestra.run.vm03.stdout:Get:40 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pastedeploy all 2.1.1-1 [26.6 kB] 2026-03-10T10:11:03.683 INFO:teuthology.orchestra.run.vm03.stdout:Get:41 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-webtest all 2.0.35-1 [28.5 kB] 2026-03-10T10:11:03.683 INFO:teuthology.orchestra.run.vm03.stdout:Get:42 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pecan all 1.3.3-4ubuntu2 [87.3 kB] 2026-03-10T10:11:03.684 INFO:teuthology.orchestra.run.vm03.stdout:Get:43 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-werkzeug all 2.0.2+dfsg1-1ubuntu0.22.04.3 [181 kB] 2026-03-10T10:11:03.687 INFO:teuthology.orchestra.run.vm03.stdout:Get:44 https://archive.ubuntu.com/ubuntu jammy/universe amd64 libfuse2 amd64 2.9.9-5ubuntu3 [90.3 kB] 2026-03-10T10:11:03.781 INFO:teuthology.orchestra.run.vm03.stdout:Get:45 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3-asyncssh all 2.5.0-1ubuntu0.1 [189 kB] 2026-03-10T10:11:03.783 INFO:teuthology.orchestra.run.vm03.stdout:Get:46 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-repoze.lru all 0.7-2 [12.1 kB] 2026-03-10T10:11:03.784 INFO:teuthology.orchestra.run.vm03.stdout:Get:47 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-routes all 2.5.1-1ubuntu1 [89.0 kB] 2026-03-10T10:11:03.784 INFO:teuthology.orchestra.run.vm03.stdout:Get:48 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn-lib amd64 0.23.2-5ubuntu6 [2058 kB] 2026-03-10T10:11:03.880 INFO:teuthology.orchestra.run.vm03.stdout:Get:49 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-joblib all 0.17.0-4ubuntu1 [204 kB] 2026-03-10T10:11:03.882 INFO:teuthology.orchestra.run.vm03.stdout:Get:50 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-threadpoolctl all 3.1.0-1 [21.3 kB] 2026-03-10T10:11:03.883 INFO:teuthology.orchestra.run.vm03.stdout:Get:51 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn all 0.23.2-5ubuntu6 [1829 kB] 2026-03-10T10:11:03.911 INFO:teuthology.orchestra.run.vm03.stdout:Get:52 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cachetools all 5.0.0-1 [9722 B] 2026-03-10T10:11:03.911 INFO:teuthology.orchestra.run.vm03.stdout:Get:53 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-rsa all 4.8-1 [28.4 kB] 2026-03-10T10:11:03.911 INFO:teuthology.orchestra.run.vm03.stdout:Get:54 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-google-auth all 1.5.1-3 [35.7 kB] 2026-03-10T10:11:03.979 INFO:teuthology.orchestra.run.vm03.stdout:Get:55 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-requests-oauthlib all 1.3.0+ds-0.1 [18.7 kB] 2026-03-10T10:11:03.979 INFO:teuthology.orchestra.run.vm03.stdout:Get:56 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-websocket all 1.2.3-1 [34.7 kB] 2026-03-10T10:11:03.984 INFO:teuthology.orchestra.run.vm03.stdout:Get:57 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-kubernetes all 12.0.1-1ubuntu1 [353 kB] 2026-03-10T10:11:03.989 INFO:teuthology.orchestra.run.vm03.stdout:Get:58 https://archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB] 2026-03-10T10:11:04.030 INFO:teuthology.orchestra.run.vm03.stdout:Get:59 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librados2 amd64 19.2.3-678-ge911bdeb-1jammy [3597 kB] 2026-03-10T10:11:04.037 INFO:teuthology.orchestra.run.vm00.stdout:Get:81 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs2 amd64 19.2.3-678-ge911bdeb-1jammy [979 kB] 2026-03-10T10:11:04.075 INFO:teuthology.orchestra.run.vm00.stdout:Get:82 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rados amd64 19.2.3-678-ge911bdeb-1jammy [357 kB] 2026-03-10T10:11:04.077 INFO:teuthology.orchestra.run.vm03.stdout:Get:60 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB] 2026-03-10T10:11:04.079 INFO:teuthology.orchestra.run.vm03.stdout:Get:61 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB] 2026-03-10T10:11:04.080 INFO:teuthology.orchestra.run.vm03.stdout:Get:62 https://archive.ubuntu.com/ubuntu jammy/main amd64 socat amd64 1.7.4.1-3ubuntu4 [349 kB] 2026-03-10T10:11:04.086 INFO:teuthology.orchestra.run.vm03.stdout:Get:63 https://archive.ubuntu.com/ubuntu jammy/universe amd64 xmlstarlet amd64 1.6.1-2.1 [265 kB] 2026-03-10T10:11:04.088 INFO:teuthology.orchestra.run.vm03.stdout:Get:64 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-socket amd64 3.0~rc1+git+ac3201d-6 [78.9 kB] 2026-03-10T10:11:04.089 INFO:teuthology.orchestra.run.vm03.stdout:Get:65 https://archive.ubuntu.com/ubuntu jammy/universe amd64 lua-sec amd64 1.0.2-1 [37.6 kB] 2026-03-10T10:11:04.092 INFO:teuthology.orchestra.run.vm09.stdout:Get:88 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rgw amd64 19.2.3-678-ge911bdeb-1jammy [112 kB] 2026-03-10T10:11:04.092 INFO:teuthology.orchestra.run.vm09.stdout:Get:89 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libradosstriper1 amd64 19.2.3-678-ge911bdeb-1jammy [470 kB] 2026-03-10T10:11:04.094 INFO:teuthology.orchestra.run.vm09.stdout:Get:90 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-common amd64 19.2.3-678-ge911bdeb-1jammy [26.5 MB] 2026-03-10T10:11:04.174 INFO:teuthology.orchestra.run.vm00.stdout:Get:83 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-argparse all 19.2.3-678-ge911bdeb-1jammy [32.9 kB] 2026-03-10T10:11:04.176 INFO:teuthology.orchestra.run.vm00.stdout:Get:84 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-cephfs amd64 19.2.3-678-ge911bdeb-1jammy [184 kB] 2026-03-10T10:11:04.176 INFO:teuthology.orchestra.run.vm03.stdout:Get:66 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 nvme-cli amd64 1.16-3ubuntu0.3 [474 kB] 2026-03-10T10:11:04.183 INFO:teuthology.orchestra.run.vm03.stdout:Get:67 https://archive.ubuntu.com/ubuntu jammy/main amd64 pkg-config amd64 0.29.2-1ubuntu3 [48.2 kB] 2026-03-10T10:11:04.183 INFO:teuthology.orchestra.run.vm03.stdout:Get:68 https://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python-asyncssh-doc all 2.5.0-1ubuntu0.1 [309 kB] 2026-03-10T10:11:04.185 INFO:teuthology.orchestra.run.vm00.stdout:Get:85 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-common all 19.2.3-678-ge911bdeb-1jammy [70.1 kB] 2026-03-10T10:11:04.188 INFO:teuthology.orchestra.run.vm03.stdout:Get:69 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-iniconfig all 1.1.1-2 [6024 B] 2026-03-10T10:11:04.188 INFO:teuthology.orchestra.run.vm00.stdout:Get:86 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rbd amd64 19.2.3-678-ge911bdeb-1jammy [334 kB] 2026-03-10T10:11:04.202 INFO:teuthology.orchestra.run.vm00.stdout:Get:87 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librgw2 amd64 19.2.3-678-ge911bdeb-1jammy [6935 kB] 2026-03-10T10:11:04.275 INFO:teuthology.orchestra.run.vm03.stdout:Get:70 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pastescript all 2.0.2-4 [54.6 kB] 2026-03-10T10:11:04.276 INFO:teuthology.orchestra.run.vm03.stdout:Get:71 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pluggy all 0.13.0-7.1 [19.0 kB] 2026-03-10T10:11:04.276 INFO:teuthology.orchestra.run.vm03.stdout:Get:72 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-psutil amd64 5.9.0-1build1 [158 kB] 2026-03-10T10:11:04.278 INFO:teuthology.orchestra.run.vm03.stdout:Get:73 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-py all 1.10.0-1 [71.9 kB] 2026-03-10T10:11:04.279 INFO:teuthology.orchestra.run.vm03.stdout:Get:74 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-pygments all 2.11.2+dfsg-2ubuntu0.1 [750 kB] 2026-03-10T10:11:04.290 INFO:teuthology.orchestra.run.vm03.stdout:Get:75 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-pyinotify all 0.9.6-1.3 [24.8 kB] 2026-03-10T10:11:04.374 INFO:teuthology.orchestra.run.vm03.stdout:Get:76 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-toml all 0.10.2-1 [16.5 kB] 2026-03-10T10:11:04.374 INFO:teuthology.orchestra.run.vm03.stdout:Get:77 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pytest all 6.2.5-1ubuntu2 [214 kB] 2026-03-10T10:11:04.377 INFO:teuthology.orchestra.run.vm03.stdout:Get:78 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplejson amd64 3.17.6-1build1 [54.7 kB] 2026-03-10T10:11:04.377 INFO:teuthology.orchestra.run.vm03.stdout:Get:79 https://archive.ubuntu.com/ubuntu jammy/universe amd64 qttranslations5-l10n all 5.15.3-1 [1983 kB] 2026-03-10T10:11:04.473 INFO:teuthology.orchestra.run.vm03.stdout:Get:80 https://archive.ubuntu.com/ubuntu jammy-updates/main amd64 smartmontools amd64 7.2-1ubuntu0.1 [583 kB] 2026-03-10T10:11:04.687 INFO:teuthology.orchestra.run.vm03.stdout:Get:81 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs2 amd64 19.2.3-678-ge911bdeb-1jammy [979 kB] 2026-03-10T10:11:04.834 INFO:teuthology.orchestra.run.vm03.stdout:Get:82 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rados amd64 19.2.3-678-ge911bdeb-1jammy [357 kB] 2026-03-10T10:11:04.866 INFO:teuthology.orchestra.run.vm03.stdout:Get:83 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-argparse all 19.2.3-678-ge911bdeb-1jammy [32.9 kB] 2026-03-10T10:11:04.869 INFO:teuthology.orchestra.run.vm03.stdout:Get:84 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-cephfs amd64 19.2.3-678-ge911bdeb-1jammy [184 kB] 2026-03-10T10:11:04.900 INFO:teuthology.orchestra.run.vm00.stdout:Get:88 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rgw amd64 19.2.3-678-ge911bdeb-1jammy [112 kB] 2026-03-10T10:11:04.903 INFO:teuthology.orchestra.run.vm00.stdout:Get:89 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libradosstriper1 amd64 19.2.3-678-ge911bdeb-1jammy [470 kB] 2026-03-10T10:11:04.967 INFO:teuthology.orchestra.run.vm03.stdout:Get:85 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-common all 19.2.3-678-ge911bdeb-1jammy [70.1 kB] 2026-03-10T10:11:04.969 INFO:teuthology.orchestra.run.vm03.stdout:Get:86 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rbd amd64 19.2.3-678-ge911bdeb-1jammy [334 kB] 2026-03-10T10:11:04.995 INFO:teuthology.orchestra.run.vm03.stdout:Get:87 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 librgw2 amd64 19.2.3-678-ge911bdeb-1jammy [6935 kB] 2026-03-10T10:11:04.996 INFO:teuthology.orchestra.run.vm00.stdout:Get:90 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-common amd64 19.2.3-678-ge911bdeb-1jammy [26.5 MB] 2026-03-10T10:11:06.257 INFO:teuthology.orchestra.run.vm03.stdout:Get:88 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 python3-rgw amd64 19.2.3-678-ge911bdeb-1jammy [112 kB] 2026-03-10T10:11:06.264 INFO:teuthology.orchestra.run.vm03.stdout:Get:89 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libradosstriper1 amd64 19.2.3-678-ge911bdeb-1jammy [470 kB] 2026-03-10T10:11:06.308 INFO:teuthology.orchestra.run.vm03.stdout:Get:90 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-common amd64 19.2.3-678-ge911bdeb-1jammy [26.5 MB] 2026-03-10T10:11:06.348 INFO:teuthology.orchestra.run.vm09.stdout:Get:91 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-base amd64 19.2.3-678-ge911bdeb-1jammy [5178 kB] 2026-03-10T10:11:06.767 INFO:teuthology.orchestra.run.vm09.stdout:Get:92 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-modules-core all 19.2.3-678-ge911bdeb-1jammy [248 kB] 2026-03-10T10:11:06.776 INFO:teuthology.orchestra.run.vm09.stdout:Get:93 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libsqlite3-mod-ceph amd64 19.2.3-678-ge911bdeb-1jammy [125 kB] 2026-03-10T10:11:06.781 INFO:teuthology.orchestra.run.vm09.stdout:Get:94 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr amd64 19.2.3-678-ge911bdeb-1jammy [1081 kB] 2026-03-10T10:11:06.891 INFO:teuthology.orchestra.run.vm09.stdout:Get:95 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mon amd64 19.2.3-678-ge911bdeb-1jammy [6239 kB] 2026-03-10T10:11:07.354 INFO:teuthology.orchestra.run.vm09.stdout:Get:96 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-osd amd64 19.2.3-678-ge911bdeb-1jammy [23.0 MB] 2026-03-10T10:11:07.568 INFO:teuthology.orchestra.run.vm00.stdout:Get:91 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-base amd64 19.2.3-678-ge911bdeb-1jammy [5178 kB] 2026-03-10T10:11:08.028 INFO:teuthology.orchestra.run.vm00.stdout:Get:92 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-modules-core all 19.2.3-678-ge911bdeb-1jammy [248 kB] 2026-03-10T10:11:08.037 INFO:teuthology.orchestra.run.vm00.stdout:Get:93 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libsqlite3-mod-ceph amd64 19.2.3-678-ge911bdeb-1jammy [125 kB] 2026-03-10T10:11:08.046 INFO:teuthology.orchestra.run.vm00.stdout:Get:94 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr amd64 19.2.3-678-ge911bdeb-1jammy [1081 kB] 2026-03-10T10:11:08.171 INFO:teuthology.orchestra.run.vm00.stdout:Get:95 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mon amd64 19.2.3-678-ge911bdeb-1jammy [6239 kB] 2026-03-10T10:11:08.779 INFO:teuthology.orchestra.run.vm00.stdout:Get:96 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-osd amd64 19.2.3-678-ge911bdeb-1jammy [23.0 MB] 2026-03-10T10:11:09.214 INFO:teuthology.orchestra.run.vm09.stdout:Get:97 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph amd64 19.2.3-678-ge911bdeb-1jammy [14.2 kB] 2026-03-10T10:11:09.214 INFO:teuthology.orchestra.run.vm09.stdout:Get:98 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-fuse amd64 19.2.3-678-ge911bdeb-1jammy [1173 kB] 2026-03-10T10:11:09.339 INFO:teuthology.orchestra.run.vm09.stdout:Get:99 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mds amd64 19.2.3-678-ge911bdeb-1jammy [2503 kB] 2026-03-10T10:11:09.603 INFO:teuthology.orchestra.run.vm09.stdout:Get:100 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 cephadm amd64 19.2.3-678-ge911bdeb-1jammy [798 kB] 2026-03-10T10:11:09.640 INFO:teuthology.orchestra.run.vm09.stdout:Get:101 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-cephadm all 19.2.3-678-ge911bdeb-1jammy [157 kB] 2026-03-10T10:11:09.646 INFO:teuthology.orchestra.run.vm09.stdout:Get:102 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-dashboard all 19.2.3-678-ge911bdeb-1jammy [2396 kB] 2026-03-10T10:11:09.903 INFO:teuthology.orchestra.run.vm09.stdout:Get:103 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-diskprediction-local all 19.2.3-678-ge911bdeb-1jammy [8625 kB] 2026-03-10T10:11:10.796 INFO:teuthology.orchestra.run.vm09.stdout:Get:104 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-k8sevents all 19.2.3-678-ge911bdeb-1jammy [14.3 kB] 2026-03-10T10:11:10.796 INFO:teuthology.orchestra.run.vm09.stdout:Get:105 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-test amd64 19.2.3-678-ge911bdeb-1jammy [52.1 MB] 2026-03-10T10:11:11.631 INFO:teuthology.orchestra.run.vm03.stdout:Get:91 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-base amd64 19.2.3-678-ge911bdeb-1jammy [5178 kB] 2026-03-10T10:11:12.390 INFO:teuthology.orchestra.run.vm00.stdout:Get:97 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph amd64 19.2.3-678-ge911bdeb-1jammy [14.2 kB] 2026-03-10T10:11:12.391 INFO:teuthology.orchestra.run.vm00.stdout:Get:98 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-fuse amd64 19.2.3-678-ge911bdeb-1jammy [1173 kB] 2026-03-10T10:11:12.698 INFO:teuthology.orchestra.run.vm00.stdout:Get:99 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mds amd64 19.2.3-678-ge911bdeb-1jammy [2503 kB] 2026-03-10T10:11:13.116 INFO:teuthology.orchestra.run.vm03.stdout:Get:92 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-modules-core all 19.2.3-678-ge911bdeb-1jammy [248 kB] 2026-03-10T10:11:13.225 INFO:teuthology.orchestra.run.vm00.stdout:Get:100 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 cephadm amd64 19.2.3-678-ge911bdeb-1jammy [798 kB] 2026-03-10T10:11:13.226 INFO:teuthology.orchestra.run.vm03.stdout:Get:93 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libsqlite3-mod-ceph amd64 19.2.3-678-ge911bdeb-1jammy [125 kB] 2026-03-10T10:11:13.283 INFO:teuthology.orchestra.run.vm03.stdout:Get:94 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr amd64 19.2.3-678-ge911bdeb-1jammy [1081 kB] 2026-03-10T10:11:13.591 INFO:teuthology.orchestra.run.vm00.stdout:Get:101 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-cephadm all 19.2.3-678-ge911bdeb-1jammy [157 kB] 2026-03-10T10:11:13.591 INFO:teuthology.orchestra.run.vm00.stdout:Get:102 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-dashboard all 19.2.3-678-ge911bdeb-1jammy [2396 kB] 2026-03-10T10:11:13.656 INFO:teuthology.orchestra.run.vm03.stdout:Get:95 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mon amd64 19.2.3-678-ge911bdeb-1jammy [6239 kB] 2026-03-10T10:11:14.123 INFO:teuthology.orchestra.run.vm00.stdout:Get:103 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-diskprediction-local all 19.2.3-678-ge911bdeb-1jammy [8625 kB] 2026-03-10T10:11:15.565 INFO:teuthology.orchestra.run.vm03.stdout:Get:96 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-osd amd64 19.2.3-678-ge911bdeb-1jammy [23.0 MB] 2026-03-10T10:11:16.206 INFO:teuthology.orchestra.run.vm00.stdout:Get:104 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-k8sevents all 19.2.3-678-ge911bdeb-1jammy [14.3 kB] 2026-03-10T10:11:16.209 INFO:teuthology.orchestra.run.vm00.stdout:Get:105 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-test amd64 19.2.3-678-ge911bdeb-1jammy [52.1 MB] 2026-03-10T10:11:19.436 INFO:teuthology.orchestra.run.vm09.stdout:Get:106 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-volume all 19.2.3-678-ge911bdeb-1jammy [135 kB] 2026-03-10T10:11:19.458 INFO:teuthology.orchestra.run.vm09.stdout:Get:107 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-dev amd64 19.2.3-678-ge911bdeb-1jammy [41.0 kB] 2026-03-10T10:11:19.463 INFO:teuthology.orchestra.run.vm09.stdout:Get:108 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 radosgw amd64 19.2.3-678-ge911bdeb-1jammy [13.7 MB] 2026-03-10T10:11:21.884 INFO:teuthology.orchestra.run.vm09.stdout:Get:109 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 rbd-fuse amd64 19.2.3-678-ge911bdeb-1jammy [92.2 kB] 2026-03-10T10:11:22.215 INFO:teuthology.orchestra.run.vm09.stdout:Fetched 178 MB in 20s (8701 kB/s) 2026-03-10T10:11:22.293 INFO:teuthology.orchestra.run.vm03.stdout:Get:97 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph amd64 19.2.3-678-ge911bdeb-1jammy [14.2 kB] 2026-03-10T10:11:22.300 INFO:teuthology.orchestra.run.vm03.stdout:Get:98 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-fuse amd64 19.2.3-678-ge911bdeb-1jammy [1173 kB] 2026-03-10T10:11:22.333 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package liblttng-ust1:amd64. 2026-03-10T10:11:22.366 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 111717 files and directories currently installed.) 2026-03-10T10:11:22.368 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../000-liblttng-ust1_2.13.1-1ubuntu1_amd64.deb ... 2026-03-10T10:11:22.370 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-10T10:11:22.390 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libdouble-conversion3:amd64. 2026-03-10T10:11:22.395 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../001-libdouble-conversion3_3.1.7-4_amd64.deb ... 2026-03-10T10:11:22.395 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-10T10:11:22.411 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libpcre2-16-0:amd64. 2026-03-10T10:11:22.417 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../002-libpcre2-16-0_10.39-3ubuntu0.1_amd64.deb ... 2026-03-10T10:11:22.418 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-10T10:11:22.439 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libqt5core5a:amd64. 2026-03-10T10:11:22.444 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../003-libqt5core5a_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-10T10:11:22.448 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:11:22.488 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libqt5dbus5:amd64. 2026-03-10T10:11:22.493 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../004-libqt5dbus5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-10T10:11:22.494 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:11:22.513 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libqt5network5:amd64. 2026-03-10T10:11:22.519 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../005-libqt5network5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-10T10:11:22.519 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:11:22.545 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libthrift-0.16.0:amd64. 2026-03-10T10:11:22.551 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../006-libthrift-0.16.0_0.16.0-2_amd64.deb ... 2026-03-10T10:11:22.551 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-10T10:11:22.575 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../007-librbd1_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:22.578 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking librbd1 (19.2.3-678-ge911bdeb-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-10T10:11:22.595 INFO:teuthology.orchestra.run.vm03.stdout:Get:99 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mds amd64 19.2.3-678-ge911bdeb-1jammy [2503 kB] 2026-03-10T10:11:22.653 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../008-librados2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:22.655 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking librados2 (19.2.3-678-ge911bdeb-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-10T10:11:22.721 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libnbd0. 2026-03-10T10:11:22.727 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../009-libnbd0_1.10.5-1_amd64.deb ... 2026-03-10T10:11:22.727 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libnbd0 (1.10.5-1) ... 2026-03-10T10:11:22.743 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libcephfs2. 2026-03-10T10:11:22.748 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../010-libcephfs2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:22.748 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:22.776 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-rados. 2026-03-10T10:11:22.781 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../011-python3-rados_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:22.782 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:22.802 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-ceph-argparse. 2026-03-10T10:11:22.807 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../012-python3-ceph-argparse_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:22.808 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:22.822 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-cephfs. 2026-03-10T10:11:22.827 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../013-python3-cephfs_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:22.828 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:22.845 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-ceph-common. 2026-03-10T10:11:22.849 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../014-python3-ceph-common_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:22.850 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:22.870 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-wcwidth. 2026-03-10T10:11:22.876 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../015-python3-wcwidth_0.2.5+dfsg1-1_all.deb ... 2026-03-10T10:11:22.877 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-10T10:11:22.896 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-prettytable. 2026-03-10T10:11:22.901 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../016-python3-prettytable_2.5.0-2_all.deb ... 2026-03-10T10:11:22.902 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-prettytable (2.5.0-2) ... 2026-03-10T10:11:22.915 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-rbd. 2026-03-10T10:11:22.919 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../017-python3-rbd_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:22.919 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:22.941 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package librdkafka1:amd64. 2026-03-10T10:11:22.947 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../018-librdkafka1_1.8.0-1build1_amd64.deb ... 2026-03-10T10:11:22.948 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-10T10:11:22.971 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libreadline-dev:amd64. 2026-03-10T10:11:22.977 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../019-libreadline-dev_8.1.2-1_amd64.deb ... 2026-03-10T10:11:22.978 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libreadline-dev:amd64 (8.1.2-1) ... 2026-03-10T10:11:22.998 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package liblua5.3-dev:amd64. 2026-03-10T10:11:23.004 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../020-liblua5.3-dev_5.3.6-1build1_amd64.deb ... 2026-03-10T10:11:23.004 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-10T10:11:23.027 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package lua5.1. 2026-03-10T10:11:23.033 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../021-lua5.1_5.1.5-8.1build4_amd64.deb ... 2026-03-10T10:11:23.033 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking lua5.1 (5.1.5-8.1build4) ... 2026-03-10T10:11:23.054 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package lua-any. 2026-03-10T10:11:23.059 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../022-lua-any_27ubuntu1_all.deb ... 2026-03-10T10:11:23.059 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking lua-any (27ubuntu1) ... 2026-03-10T10:11:23.073 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package zip. 2026-03-10T10:11:23.079 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../023-zip_3.0-12build2_amd64.deb ... 2026-03-10T10:11:23.080 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking zip (3.0-12build2) ... 2026-03-10T10:11:23.097 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package unzip. 2026-03-10T10:11:23.103 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../024-unzip_6.0-26ubuntu3.2_amd64.deb ... 2026-03-10T10:11:23.103 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking unzip (6.0-26ubuntu3.2) ... 2026-03-10T10:11:23.122 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package luarocks. 2026-03-10T10:11:23.127 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../025-luarocks_3.8.0+dfsg1-1_all.deb ... 2026-03-10T10:11:23.128 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking luarocks (3.8.0+dfsg1-1) ... 2026-03-10T10:11:23.179 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package librgw2. 2026-03-10T10:11:23.186 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../026-librgw2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:23.187 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:23.215 INFO:teuthology.orchestra.run.vm03.stdout:Get:100 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 cephadm amd64 19.2.3-678-ge911bdeb-1jammy [798 kB] 2026-03-10T10:11:23.408 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-rgw. 2026-03-10T10:11:23.409 INFO:teuthology.orchestra.run.vm03.stdout:Get:101 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-cephadm all 19.2.3-678-ge911bdeb-1jammy [157 kB] 2026-03-10T10:11:23.413 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../027-python3-rgw_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:23.414 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:23.436 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package liboath0:amd64. 2026-03-10T10:11:23.441 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../028-liboath0_2.6.7-3ubuntu0.1_amd64.deb ... 2026-03-10T10:11:23.442 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-10T10:11:23.452 INFO:teuthology.orchestra.run.vm03.stdout:Get:102 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-dashboard all 19.2.3-678-ge911bdeb-1jammy [2396 kB] 2026-03-10T10:11:23.464 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libradosstriper1. 2026-03-10T10:11:23.470 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../029-libradosstriper1_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:23.471 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:23.497 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-common. 2026-03-10T10:11:23.504 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../030-ceph-common_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:23.505 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:23.915 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-base. 2026-03-10T10:11:23.921 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../031-ceph-base_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:23.925 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:24.029 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-jaraco.functools. 2026-03-10T10:11:24.034 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../032-python3-jaraco.functools_3.4.0-2_all.deb ... 2026-03-10T10:11:24.035 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-jaraco.functools (3.4.0-2) ... 2026-03-10T10:11:24.050 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-cheroot. 2026-03-10T10:11:24.056 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../033-python3-cheroot_8.5.2+ds1-1ubuntu3.1_all.deb ... 2026-03-10T10:11:24.057 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-10T10:11:24.077 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-jaraco.classes. 2026-03-10T10:11:24.083 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../034-python3-jaraco.classes_3.2.1-3_all.deb ... 2026-03-10T10:11:24.084 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-jaraco.classes (3.2.1-3) ... 2026-03-10T10:11:24.101 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-jaraco.text. 2026-03-10T10:11:24.101 INFO:teuthology.orchestra.run.vm03.stdout:Get:103 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-diskprediction-local all 19.2.3-678-ge911bdeb-1jammy [8625 kB] 2026-03-10T10:11:24.107 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../035-python3-jaraco.text_3.6.0-2_all.deb ... 2026-03-10T10:11:24.108 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-jaraco.text (3.6.0-2) ... 2026-03-10T10:11:24.122 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-jaraco.collections. 2026-03-10T10:11:24.128 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../036-python3-jaraco.collections_3.4.0-2_all.deb ... 2026-03-10T10:11:24.129 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-jaraco.collections (3.4.0-2) ... 2026-03-10T10:11:24.144 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-tempora. 2026-03-10T10:11:24.150 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../037-python3-tempora_4.1.2-1_all.deb ... 2026-03-10T10:11:24.151 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-tempora (4.1.2-1) ... 2026-03-10T10:11:24.167 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-portend. 2026-03-10T10:11:24.173 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../038-python3-portend_3.0.0-1_all.deb ... 2026-03-10T10:11:24.174 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-portend (3.0.0-1) ... 2026-03-10T10:11:24.189 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-zc.lockfile. 2026-03-10T10:11:24.195 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../039-python3-zc.lockfile_2.0-1_all.deb ... 2026-03-10T10:11:24.196 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-zc.lockfile (2.0-1) ... 2026-03-10T10:11:24.211 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-cherrypy3. 2026-03-10T10:11:24.217 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../040-python3-cherrypy3_18.6.1-4_all.deb ... 2026-03-10T10:11:24.218 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-cherrypy3 (18.6.1-4) ... 2026-03-10T10:11:24.247 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-natsort. 2026-03-10T10:11:24.253 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../041-python3-natsort_8.0.2-1_all.deb ... 2026-03-10T10:11:24.254 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-natsort (8.0.2-1) ... 2026-03-10T10:11:24.270 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-logutils. 2026-03-10T10:11:24.275 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../042-python3-logutils_0.3.3-8_all.deb ... 2026-03-10T10:11:24.276 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-logutils (0.3.3-8) ... 2026-03-10T10:11:24.292 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-mako. 2026-03-10T10:11:24.298 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../043-python3-mako_1.1.3+ds1-2ubuntu0.1_all.deb ... 2026-03-10T10:11:24.299 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-10T10:11:24.317 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-simplegeneric. 2026-03-10T10:11:24.323 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../044-python3-simplegeneric_0.8.1-3_all.deb ... 2026-03-10T10:11:24.324 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-simplegeneric (0.8.1-3) ... 2026-03-10T10:11:24.339 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-singledispatch. 2026-03-10T10:11:24.345 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../045-python3-singledispatch_3.4.0.3-3_all.deb ... 2026-03-10T10:11:24.345 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-singledispatch (3.4.0.3-3) ... 2026-03-10T10:11:24.360 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-webob. 2026-03-10T10:11:24.365 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../046-python3-webob_1%3a1.8.6-1.1ubuntu0.1_all.deb ... 2026-03-10T10:11:24.366 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-10T10:11:24.385 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-waitress. 2026-03-10T10:11:24.390 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../047-python3-waitress_1.4.4-1.1ubuntu1.1_all.deb ... 2026-03-10T10:11:24.392 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-10T10:11:24.412 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-tempita. 2026-03-10T10:11:24.418 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../048-python3-tempita_0.5.2-6ubuntu1_all.deb ... 2026-03-10T10:11:24.419 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-tempita (0.5.2-6ubuntu1) ... 2026-03-10T10:11:24.434 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-paste. 2026-03-10T10:11:24.440 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../049-python3-paste_3.5.0+dfsg1-1_all.deb ... 2026-03-10T10:11:24.441 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-paste (3.5.0+dfsg1-1) ... 2026-03-10T10:11:24.474 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python-pastedeploy-tpl. 2026-03-10T10:11:24.480 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../050-python-pastedeploy-tpl_2.1.1-1_all.deb ... 2026-03-10T10:11:24.480 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python-pastedeploy-tpl (2.1.1-1) ... 2026-03-10T10:11:24.494 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-pastedeploy. 2026-03-10T10:11:24.499 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../051-python3-pastedeploy_2.1.1-1_all.deb ... 2026-03-10T10:11:24.500 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-pastedeploy (2.1.1-1) ... 2026-03-10T10:11:24.516 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-webtest. 2026-03-10T10:11:24.521 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../052-python3-webtest_2.0.35-1_all.deb ... 2026-03-10T10:11:24.522 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-webtest (2.0.35-1) ... 2026-03-10T10:11:24.538 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-pecan. 2026-03-10T10:11:24.544 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../053-python3-pecan_1.3.3-4ubuntu2_all.deb ... 2026-03-10T10:11:24.545 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-pecan (1.3.3-4ubuntu2) ... 2026-03-10T10:11:24.575 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-werkzeug. 2026-03-10T10:11:24.581 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../054-python3-werkzeug_2.0.2+dfsg1-1ubuntu0.22.04.3_all.deb ... 2026-03-10T10:11:24.582 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-10T10:11:24.606 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-mgr-modules-core. 2026-03-10T10:11:24.612 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../055-ceph-mgr-modules-core_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:24.613 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:24.650 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libsqlite3-mod-ceph. 2026-03-10T10:11:24.656 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../056-libsqlite3-mod-ceph_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:24.657 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:24.673 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-mgr. 2026-03-10T10:11:24.679 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../057-ceph-mgr_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:24.679 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:24.709 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-mon. 2026-03-10T10:11:24.716 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../058-ceph-mon_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:24.716 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:24.825 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libfuse2:amd64. 2026-03-10T10:11:24.831 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../059-libfuse2_2.9.9-5ubuntu3_amd64.deb ... 2026-03-10T10:11:24.832 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-10T10:11:24.851 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-osd. 2026-03-10T10:11:24.856 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../060-ceph-osd_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:24.857 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:25.226 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph. 2026-03-10T10:11:25.232 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../061-ceph_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:25.233 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:25.249 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-fuse. 2026-03-10T10:11:25.255 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../062-ceph-fuse_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:25.256 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:25.288 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-mds. 2026-03-10T10:11:25.294 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../063-ceph-mds_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:25.295 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:25.342 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package cephadm. 2026-03-10T10:11:25.348 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../064-cephadm_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:25.348 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:25.366 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-asyncssh. 2026-03-10T10:11:25.372 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../065-python3-asyncssh_2.5.0-1ubuntu0.1_all.deb ... 2026-03-10T10:11:25.373 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-10T10:11:25.399 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-mgr-cephadm. 2026-03-10T10:11:25.405 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../066-ceph-mgr-cephadm_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:25.406 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:25.430 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-repoze.lru. 2026-03-10T10:11:25.435 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../067-python3-repoze.lru_0.7-2_all.deb ... 2026-03-10T10:11:25.436 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-repoze.lru (0.7-2) ... 2026-03-10T10:11:25.452 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-routes. 2026-03-10T10:11:25.458 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../068-python3-routes_2.5.1-1ubuntu1_all.deb ... 2026-03-10T10:11:25.458 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-routes (2.5.1-1ubuntu1) ... 2026-03-10T10:11:25.482 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-mgr-dashboard. 2026-03-10T10:11:25.487 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../069-ceph-mgr-dashboard_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:25.488 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:25.800 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-sklearn-lib:amd64. 2026-03-10T10:11:25.806 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../070-python3-sklearn-lib_0.23.2-5ubuntu6_amd64.deb ... 2026-03-10T10:11:25.807 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-10T10:11:25.877 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-joblib. 2026-03-10T10:11:25.883 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../071-python3-joblib_0.17.0-4ubuntu1_all.deb ... 2026-03-10T10:11:25.884 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-joblib (0.17.0-4ubuntu1) ... 2026-03-10T10:11:25.917 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-threadpoolctl. 2026-03-10T10:11:25.923 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../072-python3-threadpoolctl_3.1.0-1_all.deb ... 2026-03-10T10:11:25.927 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-threadpoolctl (3.1.0-1) ... 2026-03-10T10:11:25.943 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-sklearn. 2026-03-10T10:11:25.949 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../073-python3-sklearn_0.23.2-5ubuntu6_all.deb ... 2026-03-10T10:11:25.949 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-10T10:11:26.073 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-mgr-diskprediction-local. 2026-03-10T10:11:26.079 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../074-ceph-mgr-diskprediction-local_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:26.080 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:26.375 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-cachetools. 2026-03-10T10:11:26.381 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../075-python3-cachetools_5.0.0-1_all.deb ... 2026-03-10T10:11:26.382 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-cachetools (5.0.0-1) ... 2026-03-10T10:11:26.397 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-rsa. 2026-03-10T10:11:26.402 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../076-python3-rsa_4.8-1_all.deb ... 2026-03-10T10:11:26.403 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-rsa (4.8-1) ... 2026-03-10T10:11:26.422 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-google-auth. 2026-03-10T10:11:26.428 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../077-python3-google-auth_1.5.1-3_all.deb ... 2026-03-10T10:11:26.429 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-google-auth (1.5.1-3) ... 2026-03-10T10:11:26.449 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-requests-oauthlib. 2026-03-10T10:11:26.457 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../078-python3-requests-oauthlib_1.3.0+ds-0.1_all.deb ... 2026-03-10T10:11:26.458 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-10T10:11:26.477 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-websocket. 2026-03-10T10:11:26.482 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../079-python3-websocket_1.2.3-1_all.deb ... 2026-03-10T10:11:26.483 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-websocket (1.2.3-1) ... 2026-03-10T10:11:26.502 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-kubernetes. 2026-03-10T10:11:26.508 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../080-python3-kubernetes_12.0.1-1ubuntu1_all.deb ... 2026-03-10T10:11:26.521 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-10T10:11:26.668 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-mgr-k8sevents. 2026-03-10T10:11:26.674 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../081-ceph-mgr-k8sevents_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:26.675 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:26.691 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libonig5:amd64. 2026-03-10T10:11:26.696 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../082-libonig5_6.9.7.1-2build1_amd64.deb ... 2026-03-10T10:11:26.697 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-10T10:11:26.716 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libjq1:amd64. 2026-03-10T10:11:26.722 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../083-libjq1_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-10T10:11:26.723 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-10T10:11:26.739 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package jq. 2026-03-10T10:11:26.745 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../084-jq_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-10T10:11:26.746 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking jq (1.6-2.1ubuntu3.1) ... 2026-03-10T10:11:26.761 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package socat. 2026-03-10T10:11:26.767 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../085-socat_1.7.4.1-3ubuntu4_amd64.deb ... 2026-03-10T10:11:26.767 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking socat (1.7.4.1-3ubuntu4) ... 2026-03-10T10:11:26.775 INFO:teuthology.orchestra.run.vm03.stdout:Get:104 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-k8sevents all 19.2.3-678-ge911bdeb-1jammy [14.3 kB] 2026-03-10T10:11:26.780 INFO:teuthology.orchestra.run.vm03.stdout:Get:105 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-test amd64 19.2.3-678-ge911bdeb-1jammy [52.1 MB] 2026-03-10T10:11:26.790 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package xmlstarlet. 2026-03-10T10:11:26.796 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../086-xmlstarlet_1.6.1-2.1_amd64.deb ... 2026-03-10T10:11:26.797 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking xmlstarlet (1.6.1-2.1) ... 2026-03-10T10:11:26.840 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-test. 2026-03-10T10:11:26.846 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../087-ceph-test_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:26.847 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:27.222 INFO:teuthology.orchestra.run.vm00.stdout:Get:106 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-volume all 19.2.3-678-ge911bdeb-1jammy [135 kB] 2026-03-10T10:11:27.223 INFO:teuthology.orchestra.run.vm00.stdout:Get:107 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-dev amd64 19.2.3-678-ge911bdeb-1jammy [41.0 kB] 2026-03-10T10:11:27.223 INFO:teuthology.orchestra.run.vm00.stdout:Get:108 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 radosgw amd64 19.2.3-678-ge911bdeb-1jammy [13.7 MB] 2026-03-10T10:11:27.690 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package ceph-volume. 2026-03-10T10:11:27.692 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../088-ceph-volume_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:27.693 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:27.721 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package libcephfs-dev. 2026-03-10T10:11:27.726 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../089-libcephfs-dev_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:27.727 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:27.743 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package lua-socket:amd64. 2026-03-10T10:11:27.748 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../090-lua-socket_3.0~rc1+git+ac3201d-6_amd64.deb ... 2026-03-10T10:11:27.749 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-10T10:11:27.774 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package lua-sec:amd64. 2026-03-10T10:11:27.779 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../091-lua-sec_1.0.2-1_amd64.deb ... 2026-03-10T10:11:27.780 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking lua-sec:amd64 (1.0.2-1) ... 2026-03-10T10:11:27.803 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package nvme-cli. 2026-03-10T10:11:27.808 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../092-nvme-cli_1.16-3ubuntu0.3_amd64.deb ... 2026-03-10T10:11:27.809 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking nvme-cli (1.16-3ubuntu0.3) ... 2026-03-10T10:11:27.848 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package pkg-config. 2026-03-10T10:11:27.853 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../093-pkg-config_0.29.2-1ubuntu3_amd64.deb ... 2026-03-10T10:11:27.854 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking pkg-config (0.29.2-1ubuntu3) ... 2026-03-10T10:11:27.868 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python-asyncssh-doc. 2026-03-10T10:11:27.873 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../094-python-asyncssh-doc_2.5.0-1ubuntu0.1_all.deb ... 2026-03-10T10:11:27.874 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-10T10:11:27.916 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-iniconfig. 2026-03-10T10:11:27.921 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../095-python3-iniconfig_1.1.1-2_all.deb ... 2026-03-10T10:11:27.922 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-iniconfig (1.1.1-2) ... 2026-03-10T10:11:27.936 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-pastescript. 2026-03-10T10:11:27.941 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../096-python3-pastescript_2.0.2-4_all.deb ... 2026-03-10T10:11:27.942 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-pastescript (2.0.2-4) ... 2026-03-10T10:11:27.961 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-pluggy. 2026-03-10T10:11:27.966 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../097-python3-pluggy_0.13.0-7.1_all.deb ... 2026-03-10T10:11:27.967 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-pluggy (0.13.0-7.1) ... 2026-03-10T10:11:27.983 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-psutil. 2026-03-10T10:11:27.988 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../098-python3-psutil_5.9.0-1build1_amd64.deb ... 2026-03-10T10:11:27.989 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-psutil (5.9.0-1build1) ... 2026-03-10T10:11:28.009 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-py. 2026-03-10T10:11:28.014 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../099-python3-py_1.10.0-1_all.deb ... 2026-03-10T10:11:28.015 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-py (1.10.0-1) ... 2026-03-10T10:11:28.036 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-pygments. 2026-03-10T10:11:28.042 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../100-python3-pygments_2.11.2+dfsg-2ubuntu0.1_all.deb ... 2026-03-10T10:11:28.042 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-10T10:11:28.100 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-pyinotify. 2026-03-10T10:11:28.106 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../101-python3-pyinotify_0.9.6-1.3_all.deb ... 2026-03-10T10:11:28.116 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-pyinotify (0.9.6-1.3) ... 2026-03-10T10:11:28.131 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-toml. 2026-03-10T10:11:28.139 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../102-python3-toml_0.10.2-1_all.deb ... 2026-03-10T10:11:28.140 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-toml (0.10.2-1) ... 2026-03-10T10:11:28.155 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-pytest. 2026-03-10T10:11:28.160 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../103-python3-pytest_6.2.5-1ubuntu2_all.deb ... 2026-03-10T10:11:28.161 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-pytest (6.2.5-1ubuntu2) ... 2026-03-10T10:11:28.188 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-simplejson. 2026-03-10T10:11:28.193 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../104-python3-simplejson_3.17.6-1build1_amd64.deb ... 2026-03-10T10:11:28.194 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-simplejson (3.17.6-1build1) ... 2026-03-10T10:11:28.212 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package qttranslations5-l10n. 2026-03-10T10:11:28.218 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../105-qttranslations5-l10n_5.15.3-1_all.deb ... 2026-03-10T10:11:28.219 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking qttranslations5-l10n (5.15.3-1) ... 2026-03-10T10:11:28.334 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package radosgw. 2026-03-10T10:11:28.340 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../106-radosgw_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:28.341 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:28.556 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package rbd-fuse. 2026-03-10T10:11:28.561 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../107-rbd-fuse_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:28.563 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:28.579 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package smartmontools. 2026-03-10T10:11:28.585 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../108-smartmontools_7.2-1ubuntu0.1_amd64.deb ... 2026-03-10T10:11:28.593 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking smartmontools (7.2-1ubuntu0.1) ... 2026-03-10T10:11:28.640 INFO:teuthology.orchestra.run.vm09.stdout:Setting up smartmontools (7.2-1ubuntu0.1) ... 2026-03-10T10:11:28.857 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/smartd.service → /lib/systemd/system/smartmontools.service. 2026-03-10T10:11:28.857 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartmontools.service → /lib/systemd/system/smartmontools.service. 2026-03-10T10:11:29.191 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-iniconfig (1.1.1-2) ... 2026-03-10T10:11:29.256 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-10T10:11:29.258 INFO:teuthology.orchestra.run.vm09.stdout:Setting up nvme-cli (1.16-3ubuntu0.3) ... 2026-03-10T10:11:29.320 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /lib/systemd/system/nvmefc-boot-connections.service. 2026-03-10T10:11:29.566 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmf-autoconnect.service → /lib/systemd/system/nvmf-autoconnect.service. 2026-03-10T10:11:29.975 INFO:teuthology.orchestra.run.vm09.stdout:nvmf-connect.target is a disabled or a static unit, not starting it. 2026-03-10T10:11:29.984 INFO:teuthology.orchestra.run.vm09.stdout:Could not execute systemctl: at /usr/bin/deb-systemd-invoke line 142. 2026-03-10T10:11:29.985 INFO:teuthology.orchestra.run.vm09.stdout:Setting up cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:30.040 INFO:teuthology.orchestra.run.vm09.stdout:Adding system user cephadm....done 2026-03-10T10:11:30.052 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-10T10:11:30.138 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-jaraco.classes (3.2.1-3) ... 2026-03-10T10:11:30.206 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-10T10:11:30.208 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-jaraco.functools (3.4.0-2) ... 2026-03-10T10:11:30.275 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-repoze.lru (0.7-2) ... 2026-03-10T10:11:30.345 INFO:teuthology.orchestra.run.vm09.stdout:Setting up liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-10T10:11:30.347 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-py (1.10.0-1) ... 2026-03-10T10:11:30.444 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-joblib (0.17.0-4ubuntu1) ... 2026-03-10T10:11:30.567 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-cachetools (5.0.0-1) ... 2026-03-10T10:11:30.635 INFO:teuthology.orchestra.run.vm09.stdout:Setting up unzip (6.0-26ubuntu3.2) ... 2026-03-10T10:11:30.644 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-pyinotify (0.9.6-1.3) ... 2026-03-10T10:11:30.710 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-threadpoolctl (3.1.0-1) ... 2026-03-10T10:11:30.776 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:30.847 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-10T10:11:30.849 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libnbd0 (1.10.5-1) ... 2026-03-10T10:11:30.851 INFO:teuthology.orchestra.run.vm09.stdout:Setting up lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-10T10:11:30.853 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libreadline-dev:amd64 (8.1.2-1) ... 2026-03-10T10:11:30.855 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-10T10:11:30.857 INFO:teuthology.orchestra.run.vm09.stdout:Setting up lua5.1 (5.1.5-8.1build4) ... 2026-03-10T10:11:30.861 INFO:teuthology.orchestra.run.vm09.stdout:update-alternatives: using /usr/bin/lua5.1 to provide /usr/bin/lua (lua-interpreter) in auto mode 2026-03-10T10:11:30.863 INFO:teuthology.orchestra.run.vm09.stdout:update-alternatives: using /usr/bin/luac5.1 to provide /usr/bin/luac (lua-compiler) in auto mode 2026-03-10T10:11:30.865 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-10T10:11:30.868 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-psutil (5.9.0-1build1) ... 2026-03-10T10:11:30.959 INFO:teuthology.orchestra.run.vm00.stdout:Get:109 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 rbd-fuse amd64 19.2.3-678-ge911bdeb-1jammy [92.2 kB] 2026-03-10T10:11:30.996 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-natsort (8.0.2-1) ... 2026-03-10T10:11:31.069 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-routes (2.5.1-1ubuntu1) ... 2026-03-10T10:11:31.147 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-simplejson (3.17.6-1build1) ... 2026-03-10T10:11:31.230 INFO:teuthology.orchestra.run.vm09.stdout:Setting up zip (3.0-12build2) ... 2026-03-10T10:11:31.233 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-10T10:11:31.344 INFO:teuthology.orchestra.run.vm00.stdout:Fetched 178 MB in 29s (6087 kB/s) 2026-03-10T10:11:31.467 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package liblttng-ust1:amd64. 2026-03-10T10:11:31.503 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 111717 files and directories currently installed.) 2026-03-10T10:11:31.505 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../000-liblttng-ust1_2.13.1-1ubuntu1_amd64.deb ... 2026-03-10T10:11:31.508 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-10T10:11:31.528 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package libdouble-conversion3:amd64. 2026-03-10T10:11:31.534 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../001-libdouble-conversion3_3.1.7-4_amd64.deb ... 2026-03-10T10:11:31.534 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-10T10:11:31.546 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-tempita (0.5.2-6ubuntu1) ... 2026-03-10T10:11:31.550 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package libpcre2-16-0:amd64. 2026-03-10T10:11:31.556 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../002-libpcre2-16-0_10.39-3ubuntu0.1_amd64.deb ... 2026-03-10T10:11:31.557 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-10T10:11:31.581 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package libqt5core5a:amd64. 2026-03-10T10:11:31.588 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../003-libqt5core5a_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-10T10:11:31.592 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:11:31.622 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python-pastedeploy-tpl (2.1.1-1) ... 2026-03-10T10:11:31.625 INFO:teuthology.orchestra.run.vm09.stdout:Setting up qttranslations5-l10n (5.15.3-1) ... 2026-03-10T10:11:31.627 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-10T10:11:31.635 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package libqt5dbus5:amd64. 2026-03-10T10:11:31.640 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../004-libqt5dbus5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-10T10:11:31.641 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:11:31.665 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package libqt5network5:amd64. 2026-03-10T10:11:31.671 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../005-libqt5network5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-10T10:11:31.672 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:11:31.699 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package libthrift-0.16.0:amd64. 2026-03-10T10:11:31.705 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../006-libthrift-0.16.0_0.16.0-2_amd64.deb ... 2026-03-10T10:11:31.705 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-10T10:11:31.716 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-10T10:11:31.731 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../007-librbd1_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:31.733 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking librbd1 (19.2.3-678-ge911bdeb-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-10T10:11:31.814 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../008-librados2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:31.817 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking librados2 (19.2.3-678-ge911bdeb-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-10T10:11:31.873 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-paste (3.5.0+dfsg1-1) ... 2026-03-10T10:11:31.890 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package libnbd0. 2026-03-10T10:11:31.896 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../009-libnbd0_1.10.5-1_amd64.deb ... 2026-03-10T10:11:31.897 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking libnbd0 (1.10.5-1) ... 2026-03-10T10:11:31.913 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package libcephfs2. 2026-03-10T10:11:31.919 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../010-libcephfs2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:31.920 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:31.948 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-rados. 2026-03-10T10:11:31.954 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../011-python3-rados_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:31.954 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:32.021 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-10T10:11:32.022 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-ceph-argparse. 2026-03-10T10:11:32.028 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../012-python3-ceph-argparse_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:32.029 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:32.047 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-cephfs. 2026-03-10T10:11:32.050 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../013-python3-cephfs_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:32.051 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:32.070 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-ceph-common. 2026-03-10T10:11:32.075 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../014-python3-ceph-common_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:32.076 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:32.100 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-wcwidth. 2026-03-10T10:11:32.107 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../015-python3-wcwidth_0.2.5+dfsg1-1_all.deb ... 2026-03-10T10:11:32.107 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-10T10:11:32.110 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-10T10:11:32.125 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-prettytable. 2026-03-10T10:11:32.131 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../016-python3-prettytable_2.5.0-2_all.deb ... 2026-03-10T10:11:32.131 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-prettytable (2.5.0-2) ... 2026-03-10T10:11:32.146 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-rbd. 2026-03-10T10:11:32.152 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../017-python3-rbd_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:32.152 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:32.173 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package librdkafka1:amd64. 2026-03-10T10:11:32.180 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../018-librdkafka1_1.8.0-1build1_amd64.deb ... 2026-03-10T10:11:32.181 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-10T10:11:32.204 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package libreadline-dev:amd64. 2026-03-10T10:11:32.211 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../019-libreadline-dev_8.1.2-1_amd64.deb ... 2026-03-10T10:11:32.212 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking libreadline-dev:amd64 (8.1.2-1) ... 2026-03-10T10:11:32.224 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-jaraco.text (3.6.0-2) ... 2026-03-10T10:11:32.231 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package liblua5.3-dev:amd64. 2026-03-10T10:11:32.236 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../020-liblua5.3-dev_5.3.6-1build1_amd64.deb ... 2026-03-10T10:11:32.262 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-10T10:11:32.283 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package lua5.1. 2026-03-10T10:11:32.288 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../021-lua5.1_5.1.5-8.1build4_amd64.deb ... 2026-03-10T10:11:32.289 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking lua5.1 (5.1.5-8.1build4) ... 2026-03-10T10:11:32.292 INFO:teuthology.orchestra.run.vm09.stdout:Setting up socat (1.7.4.1-3ubuntu4) ... 2026-03-10T10:11:32.294 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:32.308 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package lua-any. 2026-03-10T10:11:32.315 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../022-lua-any_27ubuntu1_all.deb ... 2026-03-10T10:11:32.315 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking lua-any (27ubuntu1) ... 2026-03-10T10:11:32.330 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package zip. 2026-03-10T10:11:32.336 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../023-zip_3.0-12build2_amd64.deb ... 2026-03-10T10:11:32.337 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking zip (3.0-12build2) ... 2026-03-10T10:11:32.355 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package unzip. 2026-03-10T10:11:32.360 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../024-unzip_6.0-26ubuntu3.2_amd64.deb ... 2026-03-10T10:11:32.360 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking unzip (6.0-26ubuntu3.2) ... 2026-03-10T10:11:32.379 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package luarocks. 2026-03-10T10:11:32.384 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../025-luarocks_3.8.0+dfsg1-1_all.deb ... 2026-03-10T10:11:32.385 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking luarocks (3.8.0+dfsg1-1) ... 2026-03-10T10:11:32.391 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-10T10:11:32.433 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package librgw2. 2026-03-10T10:11:32.438 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../026-librgw2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:32.439 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:32.592 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-rgw. 2026-03-10T10:11:32.599 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../027-python3-rgw_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:32.600 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:32.634 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package liboath0:amd64. 2026-03-10T10:11:32.641 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../028-liboath0_2.6.7-3ubuntu0.1_amd64.deb ... 2026-03-10T10:11:32.642 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-10T10:11:32.658 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package libradosstriper1. 2026-03-10T10:11:32.663 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../029-libradosstriper1_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:32.664 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:32.686 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package ceph-common. 2026-03-10T10:11:32.691 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../030-ceph-common_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:32.692 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:33.032 INFO:teuthology.orchestra.run.vm09.stdout:Setting up pkg-config (0.29.2-1ubuntu3) ... 2026-03-10T10:11:33.103 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package ceph-base. 2026-03-10T10:11:33.108 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:11:33.110 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../031-ceph-base_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:33.113 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-toml (0.10.2-1) ... 2026-03-10T10:11:33.115 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:33.213 INFO:teuthology.orchestra.run.vm09.stdout:Setting up librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-10T10:11:33.215 INFO:teuthology.orchestra.run.vm09.stdout:Setting up xmlstarlet (1.6.1-2.1) ... 2026-03-10T10:11:33.217 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-pluggy (0.13.0-7.1) ... 2026-03-10T10:11:33.231 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-jaraco.functools. 2026-03-10T10:11:33.239 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../032-python3-jaraco.functools_3.4.0-2_all.deb ... 2026-03-10T10:11:33.239 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-jaraco.functools (3.4.0-2) ... 2026-03-10T10:11:33.258 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-cheroot. 2026-03-10T10:11:33.264 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../033-python3-cheroot_8.5.2+ds1-1ubuntu3.1_all.deb ... 2026-03-10T10:11:33.264 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-10T10:11:33.286 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-jaraco.classes. 2026-03-10T10:11:33.288 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-zc.lockfile (2.0-1) ... 2026-03-10T10:11:33.292 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../034-python3-jaraco.classes_3.2.1-3_all.deb ... 2026-03-10T10:11:33.293 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-jaraco.classes (3.2.1-3) ... 2026-03-10T10:11:33.309 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-jaraco.text. 2026-03-10T10:11:33.315 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../035-python3-jaraco.text_3.6.0-2_all.deb ... 2026-03-10T10:11:33.315 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-jaraco.text (3.6.0-2) ... 2026-03-10T10:11:33.334 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-jaraco.collections. 2026-03-10T10:11:33.340 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../036-python3-jaraco.collections_3.4.0-2_all.deb ... 2026-03-10T10:11:33.341 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-jaraco.collections (3.4.0-2) ... 2026-03-10T10:11:33.357 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:11:33.357 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-tempora. 2026-03-10T10:11:33.359 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-rsa (4.8-1) ... 2026-03-10T10:11:33.363 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../037-python3-tempora_4.1.2-1_all.deb ... 2026-03-10T10:11:33.363 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-tempora (4.1.2-1) ... 2026-03-10T10:11:33.382 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-portend. 2026-03-10T10:11:33.388 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../038-python3-portend_3.0.0-1_all.deb ... 2026-03-10T10:11:33.388 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-portend (3.0.0-1) ... 2026-03-10T10:11:33.405 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-zc.lockfile. 2026-03-10T10:11:33.410 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../039-python3-zc.lockfile_2.0-1_all.deb ... 2026-03-10T10:11:33.411 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-zc.lockfile (2.0-1) ... 2026-03-10T10:11:33.429 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-cherrypy3. 2026-03-10T10:11:33.435 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../040-python3-cherrypy3_18.6.1-4_all.deb ... 2026-03-10T10:11:33.436 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-cherrypy3 (18.6.1-4) ... 2026-03-10T10:11:33.438 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-singledispatch (3.4.0.3-3) ... 2026-03-10T10:11:33.465 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-natsort. 2026-03-10T10:11:33.471 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../041-python3-natsort_8.0.2-1_all.deb ... 2026-03-10T10:11:33.472 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-natsort (8.0.2-1) ... 2026-03-10T10:11:33.492 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-logutils. 2026-03-10T10:11:33.499 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../042-python3-logutils_0.3.3-8_all.deb ... 2026-03-10T10:11:33.500 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-logutils (0.3.3-8) ... 2026-03-10T10:11:33.510 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-logutils (0.3.3-8) ... 2026-03-10T10:11:33.518 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-mako. 2026-03-10T10:11:33.525 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../043-python3-mako_1.1.3+ds1-2ubuntu0.1_all.deb ... 2026-03-10T10:11:33.526 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-10T10:11:33.547 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-simplegeneric. 2026-03-10T10:11:33.555 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../044-python3-simplegeneric_0.8.1-3_all.deb ... 2026-03-10T10:11:33.555 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-simplegeneric (0.8.1-3) ... 2026-03-10T10:11:33.571 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-singledispatch. 2026-03-10T10:11:33.577 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../045-python3-singledispatch_3.4.0.3-3_all.deb ... 2026-03-10T10:11:33.578 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-singledispatch (3.4.0.3-3) ... 2026-03-10T10:11:33.585 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-tempora (4.1.2-1) ... 2026-03-10T10:11:33.593 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-webob. 2026-03-10T10:11:33.599 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../046-python3-webob_1%3a1.8.6-1.1ubuntu0.1_all.deb ... 2026-03-10T10:11:33.599 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-10T10:11:33.619 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-waitress. 2026-03-10T10:11:33.624 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../047-python3-waitress_1.4.4-1.1ubuntu1.1_all.deb ... 2026-03-10T10:11:33.634 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-10T10:11:33.652 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-tempita. 2026-03-10T10:11:33.653 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-simplegeneric (0.8.1-3) ... 2026-03-10T10:11:33.658 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../048-python3-tempita_0.5.2-6ubuntu1_all.deb ... 2026-03-10T10:11:33.659 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-tempita (0.5.2-6ubuntu1) ... 2026-03-10T10:11:33.674 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-paste. 2026-03-10T10:11:33.680 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../049-python3-paste_3.5.0+dfsg1-1_all.deb ... 2026-03-10T10:11:33.681 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-paste (3.5.0+dfsg1-1) ... 2026-03-10T10:11:33.715 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python-pastedeploy-tpl. 2026-03-10T10:11:33.720 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-prettytable (2.5.0-2) ... 2026-03-10T10:11:33.721 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../050-python-pastedeploy-tpl_2.1.1-1_all.deb ... 2026-03-10T10:11:33.722 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python-pastedeploy-tpl (2.1.1-1) ... 2026-03-10T10:11:33.739 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-pastedeploy. 2026-03-10T10:11:33.744 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../051-python3-pastedeploy_2.1.1-1_all.deb ... 2026-03-10T10:11:33.745 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-pastedeploy (2.1.1-1) ... 2026-03-10T10:11:33.762 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-webtest. 2026-03-10T10:11:33.771 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../052-python3-webtest_2.0.35-1_all.deb ... 2026-03-10T10:11:33.858 INFO:teuthology.orchestra.run.vm09.stdout:Setting up liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-10T10:11:33.861 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-websocket (1.2.3-1) ... 2026-03-10T10:11:34.011 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-10T10:11:34.023 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-webtest (2.0.35-1) ... 2026-03-10T10:11:34.023 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-10T10:11:34.043 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-pecan. 2026-03-10T10:11:34.048 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../053-python3-pecan_1.3.3-4ubuntu2_all.deb ... 2026-03-10T10:11:34.049 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-pecan (1.3.3-4ubuntu2) ... 2026-03-10T10:11:34.087 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-werkzeug. 2026-03-10T10:11:34.092 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../054-python3-werkzeug_2.0.2+dfsg1-1ubuntu0.22.04.3_all.deb ... 2026-03-10T10:11:34.093 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-10T10:11:34.097 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-10T10:11:34.119 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package ceph-mgr-modules-core. 2026-03-10T10:11:34.124 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../055-ceph-mgr-modules-core_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:34.125 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:34.165 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package libsqlite3-mod-ceph. 2026-03-10T10:11:34.171 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../056-libsqlite3-mod-ceph_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:34.172 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:34.188 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-10T10:11:34.190 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package ceph-mgr. 2026-03-10T10:11:34.196 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../057-ceph-mgr_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:34.197 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:34.231 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package ceph-mon. 2026-03-10T10:11:34.236 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../058-ceph-mon_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:34.237 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:34.295 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-jaraco.collections (3.4.0-2) ... 2026-03-10T10:11:34.342 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package libfuse2:amd64. 2026-03-10T10:11:34.348 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../059-libfuse2_2.9.9-5ubuntu3_amd64.deb ... 2026-03-10T10:11:34.349 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-10T10:11:34.363 INFO:teuthology.orchestra.run.vm09.stdout:Setting up liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-10T10:11:34.366 INFO:teuthology.orchestra.run.vm09.stdout:Setting up lua-sec:amd64 (1.0.2-1) ... 2026-03-10T10:11:34.368 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package ceph-osd. 2026-03-10T10:11:34.368 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-10T10:11:34.371 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-pytest (6.2.5-1ubuntu2) ... 2026-03-10T10:11:34.373 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../060-ceph-osd_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:34.375 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:34.510 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-pastedeploy (2.1.1-1) ... 2026-03-10T10:11:34.640 INFO:teuthology.orchestra.run.vm09.stdout:Setting up lua-any (27ubuntu1) ... 2026-03-10T10:11:34.658 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-portend (3.0.0-1) ... 2026-03-10T10:11:34.670 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package ceph. 2026-03-10T10:11:34.675 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../061-ceph_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:34.676 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:34.692 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package ceph-fuse. 2026-03-10T10:11:34.697 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../062-ceph-fuse_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:34.698 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:34.724 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:11:34.727 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-google-auth (1.5.1-3) ... 2026-03-10T10:11:34.732 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package ceph-mds. 2026-03-10T10:11:34.738 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../063-ceph-mds_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:34.739 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:34.790 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package cephadm. 2026-03-10T10:11:34.795 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../064-cephadm_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:34.796 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:34.806 INFO:teuthology.orchestra.run.vm09.stdout:Setting up jq (1.6-2.1ubuntu3.1) ... 2026-03-10T10:11:34.812 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-webtest (2.0.35-1) ... 2026-03-10T10:11:34.816 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-asyncssh. 2026-03-10T10:11:34.821 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../065-python3-asyncssh_2.5.0-1ubuntu0.1_all.deb ... 2026-03-10T10:11:34.822 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-10T10:11:34.851 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package ceph-mgr-cephadm. 2026-03-10T10:11:34.857 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../066-ceph-mgr-cephadm_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:34.857 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:34.883 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-repoze.lru. 2026-03-10T10:11:34.888 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../067-python3-repoze.lru_0.7-2_all.deb ... 2026-03-10T10:11:34.890 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-repoze.lru (0.7-2) ... 2026-03-10T10:11:34.890 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-cherrypy3 (18.6.1-4) ... 2026-03-10T10:11:34.906 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-routes. 2026-03-10T10:11:34.911 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../068-python3-routes_2.5.1-1ubuntu1_all.deb ... 2026-03-10T10:11:34.912 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-routes (2.5.1-1ubuntu1) ... 2026-03-10T10:11:34.937 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package ceph-mgr-dashboard. 2026-03-10T10:11:34.943 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../069-ceph-mgr-dashboard_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:34.943 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:35.023 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-pastescript (2.0.2-4) ... 2026-03-10T10:11:35.109 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-pecan (1.3.3-4ubuntu2) ... 2026-03-10T10:11:35.260 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-10T10:11:35.309 INFO:teuthology.orchestra.run.vm09.stdout:Setting up librados2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:35.312 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:35.314 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-10T10:11:35.335 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-sklearn-lib:amd64. 2026-03-10T10:11:35.340 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../070-python3-sklearn-lib_0.23.2-5ubuntu6_amd64.deb ... 2026-03-10T10:11:35.341 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-10T10:11:35.415 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-joblib. 2026-03-10T10:11:35.421 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../071-python3-joblib_0.17.0-4ubuntu1_all.deb ... 2026-03-10T10:11:35.422 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-joblib (0.17.0-4ubuntu1) ... 2026-03-10T10:11:35.460 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-threadpoolctl. 2026-03-10T10:11:35.466 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../072-python3-threadpoolctl_3.1.0-1_all.deb ... 2026-03-10T10:11:35.467 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-threadpoolctl (3.1.0-1) ... 2026-03-10T10:11:35.486 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-sklearn. 2026-03-10T10:11:35.492 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../073-python3-sklearn_0.23.2-5ubuntu6_all.deb ... 2026-03-10T10:11:35.493 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-10T10:11:35.619 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package ceph-mgr-diskprediction-local. 2026-03-10T10:11:35.625 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../074-ceph-mgr-diskprediction-local_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:35.626 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:35.894 INFO:teuthology.orchestra.run.vm09.stdout:Setting up luarocks (3.8.0+dfsg1-1) ... 2026-03-10T10:11:35.905 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:35.907 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:35.910 INFO:teuthology.orchestra.run.vm09.stdout:Setting up librbd1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:35.912 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-cachetools. 2026-03-10T10:11:35.913 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:35.916 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:35.918 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../075-python3-cachetools_5.0.0-1_all.deb ... 2026-03-10T10:11:35.919 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-cachetools (5.0.0-1) ... 2026-03-10T10:11:35.935 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-rsa. 2026-03-10T10:11:35.940 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../076-python3-rsa_4.8-1_all.deb ... 2026-03-10T10:11:35.941 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-rsa (4.8-1) ... 2026-03-10T10:11:35.960 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-google-auth. 2026-03-10T10:11:35.965 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../077-python3-google-auth_1.5.1-3_all.deb ... 2026-03-10T10:11:35.966 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-google-auth (1.5.1-3) ... 2026-03-10T10:11:35.977 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/remote-fs.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-10T10:11:35.978 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-10T10:11:35.989 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-requests-oauthlib. 2026-03-10T10:11:35.995 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../078-python3-requests-oauthlib_1.3.0+ds-0.1_all.deb ... 2026-03-10T10:11:35.996 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-10T10:11:36.021 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-websocket. 2026-03-10T10:11:36.026 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../079-python3-websocket_1.2.3-1_all.deb ... 2026-03-10T10:11:36.027 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-websocket (1.2.3-1) ... 2026-03-10T10:11:36.050 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-kubernetes. 2026-03-10T10:11:36.055 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../080-python3-kubernetes_12.0.1-1ubuntu1_all.deb ... 2026-03-10T10:11:36.068 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-10T10:11:36.237 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package ceph-mgr-k8sevents. 2026-03-10T10:11:36.243 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../081-ceph-mgr-k8sevents_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:36.244 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:36.261 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package libonig5:amd64. 2026-03-10T10:11:36.267 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../082-libonig5_6.9.7.1-2build1_amd64.deb ... 2026-03-10T10:11:36.269 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-10T10:11:36.290 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package libjq1:amd64. 2026-03-10T10:11:36.297 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../083-libjq1_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-10T10:11:36.299 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-10T10:11:36.317 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package jq. 2026-03-10T10:11:36.323 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../084-jq_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-10T10:11:36.324 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking jq (1.6-2.1ubuntu3.1) ... 2026-03-10T10:11:36.341 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package socat. 2026-03-10T10:11:36.346 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../085-socat_1.7.4.1-3ubuntu4_amd64.deb ... 2026-03-10T10:11:36.347 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking socat (1.7.4.1-3ubuntu4) ... 2026-03-10T10:11:36.372 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package xmlstarlet. 2026-03-10T10:11:36.378 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../086-xmlstarlet_1.6.1-2.1_amd64.deb ... 2026-03-10T10:11:36.379 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking xmlstarlet (1.6.1-2.1) ... 2026-03-10T10:11:36.384 INFO:teuthology.orchestra.run.vm09.stdout:Setting up libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:36.387 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:36.389 INFO:teuthology.orchestra.run.vm09.stdout:Setting up librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:36.391 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:36.394 INFO:teuthology.orchestra.run.vm09.stdout:Setting up rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:36.396 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:36.402 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:36.407 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:36.428 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package ceph-test. 2026-03-10T10:11:36.433 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../087-ceph-test_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:36.434 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:36.445 INFO:teuthology.orchestra.run.vm09.stdout:Adding group ceph....done 2026-03-10T10:11:36.732 INFO:teuthology.orchestra.run.vm09.stdout:Adding system user ceph....done 2026-03-10T10:11:36.741 INFO:teuthology.orchestra.run.vm09.stdout:Setting system user ceph properties....done 2026-03-10T10:11:36.745 INFO:teuthology.orchestra.run.vm09.stdout:chown: cannot access '/var/log/ceph/*.log*': No such file or directory 2026-03-10T10:11:36.810 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /lib/systemd/system/ceph.target. 2026-03-10T10:11:37.040 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/rbdmap.service → /lib/systemd/system/rbdmap.service. 2026-03-10T10:11:37.301 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package ceph-volume. 2026-03-10T10:11:37.307 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../088-ceph-volume_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:37.309 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:37.341 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package libcephfs-dev. 2026-03-10T10:11:37.347 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../089-libcephfs-dev_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:37.348 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:37.354 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:37.358 INFO:teuthology.orchestra.run.vm09.stdout:Setting up radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:37.401 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package lua-socket:amd64. 2026-03-10T10:11:37.407 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../090-lua-socket_3.0~rc1+git+ac3201d-6_amd64.deb ... 2026-03-10T10:11:37.408 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-10T10:11:37.441 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package lua-sec:amd64. 2026-03-10T10:11:37.446 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../091-lua-sec_1.0.2-1_amd64.deb ... 2026-03-10T10:11:37.448 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking lua-sec:amd64 (1.0.2-1) ... 2026-03-10T10:11:37.470 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package nvme-cli. 2026-03-10T10:11:37.476 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../092-nvme-cli_1.16-3ubuntu0.3_amd64.deb ... 2026-03-10T10:11:37.477 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking nvme-cli (1.16-3ubuntu0.3) ... 2026-03-10T10:11:37.521 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package pkg-config. 2026-03-10T10:11:37.527 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../093-pkg-config_0.29.2-1ubuntu3_amd64.deb ... 2026-03-10T10:11:37.528 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking pkg-config (0.29.2-1ubuntu3) ... 2026-03-10T10:11:37.545 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python-asyncssh-doc. 2026-03-10T10:11:37.551 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../094-python-asyncssh-doc_2.5.0-1ubuntu0.1_all.deb ... 2026-03-10T10:11:37.552 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-10T10:11:37.598 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-iniconfig. 2026-03-10T10:11:37.604 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../095-python3-iniconfig_1.1.1-2_all.deb ... 2026-03-10T10:11:37.605 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-iniconfig (1.1.1-2) ... 2026-03-10T10:11:37.623 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-pastescript. 2026-03-10T10:11:37.624 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-10T10:11:37.624 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-10T10:11:37.629 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../096-python3-pastescript_2.0.2-4_all.deb ... 2026-03-10T10:11:37.630 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-pastescript (2.0.2-4) ... 2026-03-10T10:11:37.652 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-pluggy. 2026-03-10T10:11:37.658 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../097-python3-pluggy_0.13.0-7.1_all.deb ... 2026-03-10T10:11:37.659 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-pluggy (0.13.0-7.1) ... 2026-03-10T10:11:37.677 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-psutil. 2026-03-10T10:11:37.683 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../098-python3-psutil_5.9.0-1build1_amd64.deb ... 2026-03-10T10:11:37.684 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-psutil (5.9.0-1build1) ... 2026-03-10T10:11:37.708 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-py. 2026-03-10T10:11:37.715 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../099-python3-py_1.10.0-1_all.deb ... 2026-03-10T10:11:37.716 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-py (1.10.0-1) ... 2026-03-10T10:11:37.741 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-pygments. 2026-03-10T10:11:37.747 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../100-python3-pygments_2.11.2+dfsg-2ubuntu0.1_all.deb ... 2026-03-10T10:11:37.748 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-10T10:11:37.811 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-pyinotify. 2026-03-10T10:11:37.816 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../101-python3-pyinotify_0.9.6-1.3_all.deb ... 2026-03-10T10:11:37.817 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-pyinotify (0.9.6-1.3) ... 2026-03-10T10:11:37.833 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-toml. 2026-03-10T10:11:37.839 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../102-python3-toml_0.10.2-1_all.deb ... 2026-03-10T10:11:37.840 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-toml (0.10.2-1) ... 2026-03-10T10:11:37.857 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-pytest. 2026-03-10T10:11:37.863 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../103-python3-pytest_6.2.5-1ubuntu2_all.deb ... 2026-03-10T10:11:37.864 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-pytest (6.2.5-1ubuntu2) ... 2026-03-10T10:11:37.891 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-simplejson. 2026-03-10T10:11:37.897 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../104-python3-simplejson_3.17.6-1build1_amd64.deb ... 2026-03-10T10:11:37.898 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-simplejson (3.17.6-1build1) ... 2026-03-10T10:11:37.921 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package qttranslations5-l10n. 2026-03-10T10:11:37.927 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../105-qttranslations5-l10n_5.15.3-1_all.deb ... 2026-03-10T10:11:37.928 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking qttranslations5-l10n (5.15.3-1) ... 2026-03-10T10:11:38.021 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:38.041 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package radosgw. 2026-03-10T10:11:38.048 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../106-radosgw_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:38.049 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:38.104 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /lib/systemd/system/ceph-crash.service. 2026-03-10T10:11:38.313 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package rbd-fuse. 2026-03-10T10:11:38.320 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../107-rbd-fuse_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:38.322 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:38.343 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package smartmontools. 2026-03-10T10:11:38.349 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../108-smartmontools_7.2-1ubuntu0.1_amd64.deb ... 2026-03-10T10:11:38.359 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking smartmontools (7.2-1ubuntu0.1) ... 2026-03-10T10:11:38.403 INFO:teuthology.orchestra.run.vm00.stdout:Setting up smartmontools (7.2-1ubuntu0.1) ... 2026-03-10T10:11:38.462 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:38.528 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-10T10:11:38.528 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-10T10:11:38.647 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/smartd.service → /lib/systemd/system/smartmontools.service. 2026-03-10T10:11:38.647 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartmontools.service → /lib/systemd/system/smartmontools.service. 2026-03-10T10:11:38.891 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:38.951 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-10T10:11:38.952 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-10T10:11:39.011 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-iniconfig (1.1.1-2) ... 2026-03-10T10:11:39.083 INFO:teuthology.orchestra.run.vm00.stdout:Setting up libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-10T10:11:39.086 INFO:teuthology.orchestra.run.vm00.stdout:Setting up nvme-cli (1.16-3ubuntu0.3) ... 2026-03-10T10:11:39.151 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /lib/systemd/system/nvmefc-boot-connections.service. 2026-03-10T10:11:39.308 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:39.380 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmf-autoconnect.service → /lib/systemd/system/nvmf-autoconnect.service. 2026-03-10T10:11:39.388 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-10T10:11:39.388 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-10T10:11:39.747 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:39.750 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:39.763 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:39.779 INFO:teuthology.orchestra.run.vm00.stdout:nvmf-connect.target is a disabled or a static unit, not starting it. 2026-03-10T10:11:39.785 INFO:teuthology.orchestra.run.vm00.stdout:Could not execute systemctl: at /usr/bin/deb-systemd-invoke line 142. 2026-03-10T10:11:39.787 INFO:teuthology.orchestra.run.vm00.stdout:Setting up cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:39.824 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-10T10:11:39.824 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-10T10:11:39.827 INFO:teuthology.orchestra.run.vm00.stdout:Adding system user cephadm....done 2026-03-10T10:11:39.836 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-10T10:11:39.909 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-jaraco.classes (3.2.1-3) ... 2026-03-10T10:11:39.983 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-10T10:11:39.986 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-jaraco.functools (3.4.0-2) ... 2026-03-10T10:11:40.053 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-repoze.lru (0.7-2) ... 2026-03-10T10:11:40.119 INFO:teuthology.orchestra.run.vm00.stdout:Setting up liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-10T10:11:40.122 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-py (1.10.0-1) ... 2026-03-10T10:11:40.164 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:40.178 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:40.180 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:40.193 INFO:teuthology.orchestra.run.vm09.stdout:Setting up ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:40.215 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-joblib (0.17.0-4ubuntu1) ... 2026-03-10T10:11:40.310 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for mailcap (3.70+nmu1ubuntu1) ... 2026-03-10T10:11:40.318 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T10:11:40.333 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:11:40.336 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-cachetools (5.0.0-1) ... 2026-03-10T10:11:40.408 INFO:teuthology.orchestra.run.vm00.stdout:Setting up unzip (6.0-26ubuntu3.2) ... 2026-03-10T10:11:40.416 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-pyinotify (0.9.6-1.3) ... 2026-03-10T10:11:40.424 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for install-info (6.8-4build1) ... 2026-03-10T10:11:40.487 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-threadpoolctl (3.1.0-1) ... 2026-03-10T10:11:40.556 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:40.629 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-10T10:11:40.632 INFO:teuthology.orchestra.run.vm00.stdout:Setting up libnbd0 (1.10.5-1) ... 2026-03-10T10:11:40.634 INFO:teuthology.orchestra.run.vm00.stdout:Setting up lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-10T10:11:40.637 INFO:teuthology.orchestra.run.vm00.stdout:Setting up libreadline-dev:amd64 (8.1.2-1) ... 2026-03-10T10:11:40.639 INFO:teuthology.orchestra.run.vm00.stdout:Setting up libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-10T10:11:40.641 INFO:teuthology.orchestra.run.vm00.stdout:Setting up lua5.1 (5.1.5-8.1build4) ... 2026-03-10T10:11:40.646 INFO:teuthology.orchestra.run.vm00.stdout:update-alternatives: using /usr/bin/lua5.1 to provide /usr/bin/lua (lua-interpreter) in auto mode 2026-03-10T10:11:40.648 INFO:teuthology.orchestra.run.vm00.stdout:update-alternatives: using /usr/bin/luac5.1 to provide /usr/bin/luac (lua-compiler) in auto mode 2026-03-10T10:11:40.650 INFO:teuthology.orchestra.run.vm00.stdout:Setting up libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-10T10:11:40.652 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-psutil (5.9.0-1build1) ... 2026-03-10T10:11:40.762 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:11:40.762 INFO:teuthology.orchestra.run.vm09.stdout:Running kernel seems to be up-to-date. 2026-03-10T10:11:40.762 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:11:40.762 INFO:teuthology.orchestra.run.vm09.stdout:Services to be restarted: 2026-03-10T10:11:40.768 INFO:teuthology.orchestra.run.vm09.stdout: systemctl restart packagekit.service 2026-03-10T10:11:40.771 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:11:40.771 INFO:teuthology.orchestra.run.vm09.stdout:Service restarts being deferred: 2026-03-10T10:11:40.771 INFO:teuthology.orchestra.run.vm09.stdout: systemctl restart networkd-dispatcher.service 2026-03-10T10:11:40.771 INFO:teuthology.orchestra.run.vm09.stdout: systemctl restart unattended-upgrades.service 2026-03-10T10:11:40.771 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:11:40.771 INFO:teuthology.orchestra.run.vm09.stdout:No containers need to be restarted. 2026-03-10T10:11:40.771 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:11:40.771 INFO:teuthology.orchestra.run.vm09.stdout:No user sessions are running outdated binaries. 2026-03-10T10:11:40.771 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:11:40.771 INFO:teuthology.orchestra.run.vm09.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-10T10:11:40.774 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-natsort (8.0.2-1) ... 2026-03-10T10:11:40.845 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-routes (2.5.1-1ubuntu1) ... 2026-03-10T10:11:40.916 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-simplejson (3.17.6-1build1) ... 2026-03-10T10:11:40.995 INFO:teuthology.orchestra.run.vm00.stdout:Setting up zip (3.0-12build2) ... 2026-03-10T10:11:40.997 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-10T10:11:41.282 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-tempita (0.5.2-6ubuntu1) ... 2026-03-10T10:11:41.352 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python-pastedeploy-tpl (2.1.1-1) ... 2026-03-10T10:11:41.355 INFO:teuthology.orchestra.run.vm00.stdout:Setting up qttranslations5-l10n (5.15.3-1) ... 2026-03-10T10:11:41.358 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-10T10:11:41.450 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-10T10:11:41.584 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-paste (3.5.0+dfsg1-1) ... 2026-03-10T10:11:41.668 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:11:41.670 DEBUG:teuthology.orchestra.run.vm09:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install python3-xmltodict python3-jmespath 2026-03-10T10:11:41.708 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-10T10:11:41.747 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:11:41.793 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-10T10:11:41.906 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-jaraco.text (3.6.0-2) ... 2026-03-10T10:11:41.958 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:11:41.959 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:11:41.973 INFO:teuthology.orchestra.run.vm00.stdout:Setting up socat (1.7.4.1-3ubuntu4) ... 2026-03-10T10:11:41.975 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:42.069 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-10T10:11:42.134 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:11:42.134 INFO:teuthology.orchestra.run.vm09.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T10:11:42.134 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-10T10:11:42.135 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:11:42.150 INFO:teuthology.orchestra.run.vm09.stdout:The following NEW packages will be installed: 2026-03-10T10:11:42.150 INFO:teuthology.orchestra.run.vm09.stdout: python3-jmespath python3-xmltodict 2026-03-10T10:11:42.400 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 2 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:11:42.400 INFO:teuthology.orchestra.run.vm09.stdout:Need to get 34.3 kB of archives. 2026-03-10T10:11:42.400 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 146 kB of additional disk space will be used. 2026-03-10T10:11:42.400 INFO:teuthology.orchestra.run.vm09.stdout:Get:1 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jmespath all 0.10.0-1 [21.7 kB] 2026-03-10T10:11:42.507 INFO:teuthology.orchestra.run.vm09.stdout:Get:2 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-xmltodict all 0.12.0-2 [12.6 kB] 2026-03-10T10:11:42.631 INFO:teuthology.orchestra.run.vm00.stdout:Setting up pkg-config (0.29.2-1ubuntu3) ... 2026-03-10T10:11:42.653 INFO:teuthology.orchestra.run.vm00.stdout:Setting up libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:11:42.658 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-toml (0.10.2-1) ... 2026-03-10T10:11:42.702 INFO:teuthology.orchestra.run.vm09.stdout:Fetched 34.3 kB in 0s (95.8 kB/s) 2026-03-10T10:11:42.729 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-jmespath. 2026-03-10T10:11:42.734 INFO:teuthology.orchestra.run.vm00.stdout:Setting up librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-10T10:11:42.737 INFO:teuthology.orchestra.run.vm00.stdout:Setting up xmlstarlet (1.6.1-2.1) ... 2026-03-10T10:11:42.739 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-pluggy (0.13.0-7.1) ... 2026-03-10T10:11:42.756 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118577 files and directories currently installed.) 2026-03-10T10:11:42.758 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../python3-jmespath_0.10.0-1_all.deb ... 2026-03-10T10:11:42.759 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-jmespath (0.10.0-1) ... 2026-03-10T10:11:42.776 INFO:teuthology.orchestra.run.vm09.stdout:Selecting previously unselected package python3-xmltodict. 2026-03-10T10:11:42.783 INFO:teuthology.orchestra.run.vm09.stdout:Preparing to unpack .../python3-xmltodict_0.12.0-2_all.deb ... 2026-03-10T10:11:42.784 INFO:teuthology.orchestra.run.vm09.stdout:Unpacking python3-xmltodict (0.12.0-2) ... 2026-03-10T10:11:42.814 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-zc.lockfile (2.0-1) ... 2026-03-10T10:11:42.826 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-xmltodict (0.12.0-2) ... 2026-03-10T10:11:42.881 INFO:teuthology.orchestra.run.vm00.stdout:Setting up libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:11:42.884 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-rsa (4.8-1) ... 2026-03-10T10:11:42.893 INFO:teuthology.orchestra.run.vm09.stdout:Setting up python3-jmespath (0.10.0-1) ... 2026-03-10T10:11:42.959 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-singledispatch (3.4.0.3-3) ... 2026-03-10T10:11:43.031 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-logutils (0.3.3-8) ... 2026-03-10T10:11:43.102 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-tempora (4.1.2-1) ... 2026-03-10T10:11:43.174 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-simplegeneric (0.8.1-3) ... 2026-03-10T10:11:43.234 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:11:43.234 INFO:teuthology.orchestra.run.vm09.stdout:Running kernel seems to be up-to-date. 2026-03-10T10:11:43.234 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:11:43.234 INFO:teuthology.orchestra.run.vm09.stdout:Services to be restarted: 2026-03-10T10:11:43.238 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-prettytable (2.5.0-2) ... 2026-03-10T10:11:43.240 INFO:teuthology.orchestra.run.vm09.stdout: systemctl restart packagekit.service 2026-03-10T10:11:43.243 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:11:43.243 INFO:teuthology.orchestra.run.vm09.stdout:Service restarts being deferred: 2026-03-10T10:11:43.243 INFO:teuthology.orchestra.run.vm09.stdout: systemctl restart networkd-dispatcher.service 2026-03-10T10:11:43.243 INFO:teuthology.orchestra.run.vm09.stdout: systemctl restart unattended-upgrades.service 2026-03-10T10:11:43.243 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:11:43.243 INFO:teuthology.orchestra.run.vm09.stdout:No containers need to be restarted. 2026-03-10T10:11:43.243 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:11:43.243 INFO:teuthology.orchestra.run.vm09.stdout:No user sessions are running outdated binaries. 2026-03-10T10:11:43.243 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:11:43.243 INFO:teuthology.orchestra.run.vm09.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-10T10:11:43.312 INFO:teuthology.orchestra.run.vm00.stdout:Setting up liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-10T10:11:43.314 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-websocket (1.2.3-1) ... 2026-03-10T10:11:43.397 INFO:teuthology.orchestra.run.vm00.stdout:Setting up libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-10T10:11:43.400 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-10T10:11:43.472 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-10T10:11:43.556 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-10T10:11:43.647 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-jaraco.collections (3.4.0-2) ... 2026-03-10T10:11:43.713 INFO:teuthology.orchestra.run.vm00.stdout:Setting up liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-10T10:11:43.715 INFO:teuthology.orchestra.run.vm00.stdout:Setting up lua-sec:amd64 (1.0.2-1) ... 2026-03-10T10:11:43.717 INFO:teuthology.orchestra.run.vm00.stdout:Setting up libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-10T10:11:43.719 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-pytest (6.2.5-1ubuntu2) ... 2026-03-10T10:11:43.861 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-pastedeploy (2.1.1-1) ... 2026-03-10T10:11:43.932 INFO:teuthology.orchestra.run.vm00.stdout:Setting up lua-any (27ubuntu1) ... 2026-03-10T10:11:43.935 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-portend (3.0.0-1) ... 2026-03-10T10:11:44.001 INFO:teuthology.orchestra.run.vm00.stdout:Setting up libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:11:44.003 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-google-auth (1.5.1-3) ... 2026-03-10T10:11:44.091 INFO:teuthology.orchestra.run.vm00.stdout:Setting up jq (1.6-2.1ubuntu3.1) ... 2026-03-10T10:11:44.093 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-webtest (2.0.35-1) ... 2026-03-10T10:11:44.165 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:11:44.167 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-cherrypy3 (18.6.1-4) ... 2026-03-10T10:11:44.169 DEBUG:teuthology.parallel:result is None 2026-03-10T10:11:44.291 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-pastescript (2.0.2-4) ... 2026-03-10T10:11:44.372 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-pecan (1.3.3-4ubuntu2) ... 2026-03-10T10:11:44.483 INFO:teuthology.orchestra.run.vm00.stdout:Setting up libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-10T10:11:44.485 INFO:teuthology.orchestra.run.vm00.stdout:Setting up librados2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:44.487 INFO:teuthology.orchestra.run.vm00.stdout:Setting up libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:44.489 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-10T10:11:45.071 INFO:teuthology.orchestra.run.vm00.stdout:Setting up luarocks (3.8.0+dfsg1-1) ... 2026-03-10T10:11:45.078 INFO:teuthology.orchestra.run.vm00.stdout:Setting up libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:45.080 INFO:teuthology.orchestra.run.vm00.stdout:Setting up libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:45.082 INFO:teuthology.orchestra.run.vm00.stdout:Setting up librbd1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:45.085 INFO:teuthology.orchestra.run.vm00.stdout:Setting up ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:45.087 INFO:teuthology.orchestra.run.vm00.stdout:Setting up ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:45.145 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/remote-fs.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-10T10:11:45.145 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-10T10:11:45.473 INFO:teuthology.orchestra.run.vm00.stdout:Setting up libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:45.475 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:45.477 INFO:teuthology.orchestra.run.vm00.stdout:Setting up librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:45.479 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:45.481 INFO:teuthology.orchestra.run.vm00.stdout:Setting up rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:45.484 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:45.486 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:45.488 INFO:teuthology.orchestra.run.vm00.stdout:Setting up ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:45.519 INFO:teuthology.orchestra.run.vm00.stdout:Adding group ceph....done 2026-03-10T10:11:45.553 INFO:teuthology.orchestra.run.vm00.stdout:Adding system user ceph....done 2026-03-10T10:11:45.560 INFO:teuthology.orchestra.run.vm00.stdout:Setting system user ceph properties....done 2026-03-10T10:11:45.564 INFO:teuthology.orchestra.run.vm00.stdout:chown: cannot access '/var/log/ceph/*.log*': No such file or directory 2026-03-10T10:11:45.627 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /lib/systemd/system/ceph.target. 2026-03-10T10:11:45.848 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/rbdmap.service → /lib/systemd/system/rbdmap.service. 2026-03-10T10:11:46.186 INFO:teuthology.orchestra.run.vm00.stdout:Setting up ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:46.188 INFO:teuthology.orchestra.run.vm00.stdout:Setting up radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:46.435 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-10T10:11:46.435 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-10T10:11:46.809 INFO:teuthology.orchestra.run.vm00.stdout:Setting up ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:46.891 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /lib/systemd/system/ceph-crash.service. 2026-03-10T10:11:47.249 INFO:teuthology.orchestra.run.vm00.stdout:Setting up ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:47.311 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-10T10:11:47.311 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-10T10:11:47.649 INFO:teuthology.orchestra.run.vm00.stdout:Setting up ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:47.710 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-10T10:11:47.710 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-10T10:11:48.077 INFO:teuthology.orchestra.run.vm00.stdout:Setting up ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:48.085 INFO:teuthology.orchestra.run.vm03.stdout:Get:106 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 ceph-volume all 19.2.3-678-ge911bdeb-1jammy [135 kB] 2026-03-10T10:11:48.102 INFO:teuthology.orchestra.run.vm03.stdout:Get:107 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-dev amd64 19.2.3-678-ge911bdeb-1jammy [41.0 kB] 2026-03-10T10:11:48.109 INFO:teuthology.orchestra.run.vm03.stdout:Get:108 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 radosgw amd64 19.2.3-678-ge911bdeb-1jammy [13.7 MB] 2026-03-10T10:11:48.155 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-10T10:11:48.155 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-10T10:11:48.521 INFO:teuthology.orchestra.run.vm00.stdout:Setting up ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:48.524 INFO:teuthology.orchestra.run.vm00.stdout:Setting up ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:48.540 INFO:teuthology.orchestra.run.vm00.stdout:Setting up ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:48.602 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-10T10:11:48.602 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-10T10:11:48.958 INFO:teuthology.orchestra.run.vm00.stdout:Setting up ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:48.970 INFO:teuthology.orchestra.run.vm00.stdout:Setting up ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:48.973 INFO:teuthology.orchestra.run.vm00.stdout:Setting up ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:48.985 INFO:teuthology.orchestra.run.vm00.stdout:Setting up ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:49.099 INFO:teuthology.orchestra.run.vm00.stdout:Processing triggers for mailcap (3.70+nmu1ubuntu1) ... 2026-03-10T10:11:49.107 INFO:teuthology.orchestra.run.vm00.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T10:11:49.122 INFO:teuthology.orchestra.run.vm00.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:11:49.198 INFO:teuthology.orchestra.run.vm00.stdout:Processing triggers for install-info (6.8-4build1) ... 2026-03-10T10:11:49.510 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:11:49.510 INFO:teuthology.orchestra.run.vm00.stdout:Running kernel seems to be up-to-date. 2026-03-10T10:11:49.510 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:11:49.510 INFO:teuthology.orchestra.run.vm00.stdout:Services to be restarted: 2026-03-10T10:11:49.515 INFO:teuthology.orchestra.run.vm00.stdout: systemctl restart packagekit.service 2026-03-10T10:11:49.518 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:11:49.518 INFO:teuthology.orchestra.run.vm00.stdout:Service restarts being deferred: 2026-03-10T10:11:49.518 INFO:teuthology.orchestra.run.vm00.stdout: systemctl restart networkd-dispatcher.service 2026-03-10T10:11:49.518 INFO:teuthology.orchestra.run.vm00.stdout: systemctl restart unattended-upgrades.service 2026-03-10T10:11:49.518 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:11:49.518 INFO:teuthology.orchestra.run.vm00.stdout:No containers need to be restarted. 2026-03-10T10:11:49.519 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:11:49.519 INFO:teuthology.orchestra.run.vm00.stdout:No user sessions are running outdated binaries. 2026-03-10T10:11:49.519 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:11:49.519 INFO:teuthology.orchestra.run.vm00.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-10T10:11:49.674 INFO:teuthology.orchestra.run.vm03.stdout:Get:109 https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default jammy/main amd64 rbd-fuse amd64 19.2.3-678-ge911bdeb-1jammy [92.2 kB] 2026-03-10T10:11:49.988 INFO:teuthology.orchestra.run.vm03.stdout:Fetched 178 MB in 48s (3718 kB/s) 2026-03-10T10:11:50.106 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package liblttng-ust1:amd64. 2026-03-10T10:11:50.137 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 111717 files and directories currently installed.) 2026-03-10T10:11:50.139 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../000-liblttng-ust1_2.13.1-1ubuntu1_amd64.deb ... 2026-03-10T10:11:50.141 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-10T10:11:50.162 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package libdouble-conversion3:amd64. 2026-03-10T10:11:50.168 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../001-libdouble-conversion3_3.1.7-4_amd64.deb ... 2026-03-10T10:11:50.168 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-10T10:11:50.185 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package libpcre2-16-0:amd64. 2026-03-10T10:11:50.190 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../002-libpcre2-16-0_10.39-3ubuntu0.1_amd64.deb ... 2026-03-10T10:11:50.190 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-10T10:11:50.209 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package libqt5core5a:amd64. 2026-03-10T10:11:50.214 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../003-libqt5core5a_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-10T10:11:50.217 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:11:50.260 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package libqt5dbus5:amd64. 2026-03-10T10:11:50.265 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../004-libqt5dbus5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-10T10:11:50.265 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:11:50.282 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package libqt5network5:amd64. 2026-03-10T10:11:50.285 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../005-libqt5network5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-10T10:11:50.286 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:11:50.310 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package libthrift-0.16.0:amd64. 2026-03-10T10:11:50.313 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../006-libthrift-0.16.0_0.16.0-2_amd64.deb ... 2026-03-10T10:11:50.314 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-10T10:11:50.334 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../007-librbd1_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:50.336 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking librbd1 (19.2.3-678-ge911bdeb-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-10T10:11:50.407 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../008-librados2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:50.409 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking librados2 (19.2.3-678-ge911bdeb-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-10T10:11:50.428 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:11:50.430 DEBUG:teuthology.orchestra.run.vm00:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install python3-xmltodict python3-jmespath 2026-03-10T10:11:50.476 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package libnbd0. 2026-03-10T10:11:50.483 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../009-libnbd0_1.10.5-1_amd64.deb ... 2026-03-10T10:11:50.483 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking libnbd0 (1.10.5-1) ... 2026-03-10T10:11:50.500 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package libcephfs2. 2026-03-10T10:11:50.506 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:11:50.506 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../010-libcephfs2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:50.507 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:50.536 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-rados. 2026-03-10T10:11:50.542 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../011-python3-rados_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:50.542 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:50.564 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-ceph-argparse. 2026-03-10T10:11:50.569 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../012-python3-ceph-argparse_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:50.570 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:50.586 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-cephfs. 2026-03-10T10:11:50.591 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../013-python3-cephfs_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:50.592 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:50.609 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-ceph-common. 2026-03-10T10:11:50.611 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:11:50.611 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:11:50.614 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../014-python3-ceph-common_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:50.614 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:50.634 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-wcwidth. 2026-03-10T10:11:50.640 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../015-python3-wcwidth_0.2.5+dfsg1-1_all.deb ... 2026-03-10T10:11:50.640 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-10T10:11:50.658 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-prettytable. 2026-03-10T10:11:50.663 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../016-python3-prettytable_2.5.0-2_all.deb ... 2026-03-10T10:11:50.664 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-prettytable (2.5.0-2) ... 2026-03-10T10:11:50.678 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-rbd. 2026-03-10T10:11:50.683 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../017-python3-rbd_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:50.684 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:50.704 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package librdkafka1:amd64. 2026-03-10T10:11:50.709 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../018-librdkafka1_1.8.0-1build1_amd64.deb ... 2026-03-10T10:11:50.710 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-10T10:11:50.732 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package libreadline-dev:amd64. 2026-03-10T10:11:50.737 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../019-libreadline-dev_8.1.2-1_amd64.deb ... 2026-03-10T10:11:50.738 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking libreadline-dev:amd64 (8.1.2-1) ... 2026-03-10T10:11:50.756 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package liblua5.3-dev:amd64. 2026-03-10T10:11:50.763 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../020-liblua5.3-dev_5.3.6-1build1_amd64.deb ... 2026-03-10T10:11:50.764 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-10T10:11:50.784 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package lua5.1. 2026-03-10T10:11:50.790 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../021-lua5.1_5.1.5-8.1build4_amd64.deb ... 2026-03-10T10:11:50.791 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking lua5.1 (5.1.5-8.1build4) ... 2026-03-10T10:11:50.804 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:11:50.804 INFO:teuthology.orchestra.run.vm00.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T10:11:50.805 INFO:teuthology.orchestra.run.vm00.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-10T10:11:50.805 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:11:50.810 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package lua-any. 2026-03-10T10:11:50.816 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../022-lua-any_27ubuntu1_all.deb ... 2026-03-10T10:11:50.816 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking lua-any (27ubuntu1) ... 2026-03-10T10:11:50.822 INFO:teuthology.orchestra.run.vm00.stdout:The following NEW packages will be installed: 2026-03-10T10:11:50.822 INFO:teuthology.orchestra.run.vm00.stdout: python3-jmespath python3-xmltodict 2026-03-10T10:11:50.830 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package zip. 2026-03-10T10:11:50.835 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../023-zip_3.0-12build2_amd64.deb ... 2026-03-10T10:11:50.836 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking zip (3.0-12build2) ... 2026-03-10T10:11:50.854 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package unzip. 2026-03-10T10:11:50.859 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../024-unzip_6.0-26ubuntu3.2_amd64.deb ... 2026-03-10T10:11:50.860 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking unzip (6.0-26ubuntu3.2) ... 2026-03-10T10:11:50.878 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package luarocks. 2026-03-10T10:11:50.883 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../025-luarocks_3.8.0+dfsg1-1_all.deb ... 2026-03-10T10:11:50.884 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking luarocks (3.8.0+dfsg1-1) ... 2026-03-10T10:11:50.935 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package librgw2. 2026-03-10T10:11:50.941 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../026-librgw2_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:50.941 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:51.026 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 2 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:11:51.026 INFO:teuthology.orchestra.run.vm00.stdout:Need to get 34.3 kB of archives. 2026-03-10T10:11:51.026 INFO:teuthology.orchestra.run.vm00.stdout:After this operation, 146 kB of additional disk space will be used. 2026-03-10T10:11:51.026 INFO:teuthology.orchestra.run.vm00.stdout:Get:1 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jmespath all 0.10.0-1 [21.7 kB] 2026-03-10T10:11:51.071 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-rgw. 2026-03-10T10:11:51.076 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../027-python3-rgw_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:51.077 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:51.097 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package liboath0:amd64. 2026-03-10T10:11:51.100 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../028-liboath0_2.6.7-3ubuntu0.1_amd64.deb ... 2026-03-10T10:11:51.101 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-10T10:11:51.104 INFO:teuthology.orchestra.run.vm00.stdout:Get:2 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-xmltodict all 0.12.0-2 [12.6 kB] 2026-03-10T10:11:51.119 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package libradosstriper1. 2026-03-10T10:11:51.123 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../029-libradosstriper1_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:51.124 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:51.148 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package ceph-common. 2026-03-10T10:11:51.152 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../030-ceph-common_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:51.154 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:51.315 INFO:teuthology.orchestra.run.vm00.stdout:Fetched 34.3 kB in 0s (122 kB/s) 2026-03-10T10:11:51.330 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-jmespath. 2026-03-10T10:11:51.353 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118577 files and directories currently installed.) 2026-03-10T10:11:51.355 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../python3-jmespath_0.10.0-1_all.deb ... 2026-03-10T10:11:51.356 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-jmespath (0.10.0-1) ... 2026-03-10T10:11:51.372 INFO:teuthology.orchestra.run.vm00.stdout:Selecting previously unselected package python3-xmltodict. 2026-03-10T10:11:51.377 INFO:teuthology.orchestra.run.vm00.stdout:Preparing to unpack .../python3-xmltodict_0.12.0-2_all.deb ... 2026-03-10T10:11:51.378 INFO:teuthology.orchestra.run.vm00.stdout:Unpacking python3-xmltodict (0.12.0-2) ... 2026-03-10T10:11:51.405 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-xmltodict (0.12.0-2) ... 2026-03-10T10:11:51.564 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package ceph-base. 2026-03-10T10:11:51.569 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../031-ceph-base_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:51.573 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:51.607 INFO:teuthology.orchestra.run.vm00.stdout:Setting up python3-jmespath (0.10.0-1) ... 2026-03-10T10:11:51.666 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-jaraco.functools. 2026-03-10T10:11:51.668 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../032-python3-jaraco.functools_3.4.0-2_all.deb ... 2026-03-10T10:11:51.669 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-jaraco.functools (3.4.0-2) ... 2026-03-10T10:11:51.686 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-cheroot. 2026-03-10T10:11:51.689 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../033-python3-cheroot_8.5.2+ds1-1ubuntu3.1_all.deb ... 2026-03-10T10:11:51.690 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-10T10:11:51.708 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-jaraco.classes. 2026-03-10T10:11:51.714 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../034-python3-jaraco.classes_3.2.1-3_all.deb ... 2026-03-10T10:11:51.715 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-jaraco.classes (3.2.1-3) ... 2026-03-10T10:11:51.731 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-jaraco.text. 2026-03-10T10:11:51.737 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../035-python3-jaraco.text_3.6.0-2_all.deb ... 2026-03-10T10:11:51.738 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-jaraco.text (3.6.0-2) ... 2026-03-10T10:11:51.757 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-jaraco.collections. 2026-03-10T10:11:51.761 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../036-python3-jaraco.collections_3.4.0-2_all.deb ... 2026-03-10T10:11:51.762 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-jaraco.collections (3.4.0-2) ... 2026-03-10T10:11:51.776 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-tempora. 2026-03-10T10:11:51.781 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../037-python3-tempora_4.1.2-1_all.deb ... 2026-03-10T10:11:51.782 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-tempora (4.1.2-1) ... 2026-03-10T10:11:51.798 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-portend. 2026-03-10T10:11:51.803 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../038-python3-portend_3.0.0-1_all.deb ... 2026-03-10T10:11:51.804 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-portend (3.0.0-1) ... 2026-03-10T10:11:51.820 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-zc.lockfile. 2026-03-10T10:11:51.825 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../039-python3-zc.lockfile_2.0-1_all.deb ... 2026-03-10T10:11:51.826 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-zc.lockfile (2.0-1) ... 2026-03-10T10:11:51.840 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-cherrypy3. 2026-03-10T10:11:51.846 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../040-python3-cherrypy3_18.6.1-4_all.deb ... 2026-03-10T10:11:51.847 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-cherrypy3 (18.6.1-4) ... 2026-03-10T10:11:51.879 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-natsort. 2026-03-10T10:11:51.884 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../041-python3-natsort_8.0.2-1_all.deb ... 2026-03-10T10:11:51.885 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-natsort (8.0.2-1) ... 2026-03-10T10:11:51.904 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-logutils. 2026-03-10T10:11:51.907 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../042-python3-logutils_0.3.3-8_all.deb ... 2026-03-10T10:11:51.908 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-logutils (0.3.3-8) ... 2026-03-10T10:11:51.924 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-mako. 2026-03-10T10:11:51.930 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../043-python3-mako_1.1.3+ds1-2ubuntu0.1_all.deb ... 2026-03-10T10:11:51.930 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-10T10:11:51.935 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:11:51.935 INFO:teuthology.orchestra.run.vm00.stdout:Running kernel seems to be up-to-date. 2026-03-10T10:11:51.935 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:11:51.935 INFO:teuthology.orchestra.run.vm00.stdout:Services to be restarted: 2026-03-10T10:11:51.941 INFO:teuthology.orchestra.run.vm00.stdout: systemctl restart packagekit.service 2026-03-10T10:11:51.943 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:11:51.943 INFO:teuthology.orchestra.run.vm00.stdout:Service restarts being deferred: 2026-03-10T10:11:51.943 INFO:teuthology.orchestra.run.vm00.stdout: systemctl restart networkd-dispatcher.service 2026-03-10T10:11:51.943 INFO:teuthology.orchestra.run.vm00.stdout: systemctl restart unattended-upgrades.service 2026-03-10T10:11:51.943 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:11:51.944 INFO:teuthology.orchestra.run.vm00.stdout:No containers need to be restarted. 2026-03-10T10:11:51.944 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:11:51.944 INFO:teuthology.orchestra.run.vm00.stdout:No user sessions are running outdated binaries. 2026-03-10T10:11:51.944 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:11:51.944 INFO:teuthology.orchestra.run.vm00.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-10T10:11:51.949 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-simplegeneric. 2026-03-10T10:11:51.955 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../044-python3-simplegeneric_0.8.1-3_all.deb ... 2026-03-10T10:11:51.955 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-simplegeneric (0.8.1-3) ... 2026-03-10T10:11:51.969 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-singledispatch. 2026-03-10T10:11:51.974 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../045-python3-singledispatch_3.4.0.3-3_all.deb ... 2026-03-10T10:11:51.975 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-singledispatch (3.4.0.3-3) ... 2026-03-10T10:11:51.989 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-webob. 2026-03-10T10:11:51.994 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../046-python3-webob_1%3a1.8.6-1.1ubuntu0.1_all.deb ... 2026-03-10T10:11:51.995 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-10T10:11:52.014 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-waitress. 2026-03-10T10:11:52.020 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../047-python3-waitress_1.4.4-1.1ubuntu1.1_all.deb ... 2026-03-10T10:11:52.022 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-10T10:11:52.041 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-tempita. 2026-03-10T10:11:52.047 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../048-python3-tempita_0.5.2-6ubuntu1_all.deb ... 2026-03-10T10:11:52.048 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-tempita (0.5.2-6ubuntu1) ... 2026-03-10T10:11:52.063 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-paste. 2026-03-10T10:11:52.068 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../049-python3-paste_3.5.0+dfsg1-1_all.deb ... 2026-03-10T10:11:52.069 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-paste (3.5.0+dfsg1-1) ... 2026-03-10T10:11:52.101 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python-pastedeploy-tpl. 2026-03-10T10:11:52.108 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../050-python-pastedeploy-tpl_2.1.1-1_all.deb ... 2026-03-10T10:11:52.109 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python-pastedeploy-tpl (2.1.1-1) ... 2026-03-10T10:11:52.124 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-pastedeploy. 2026-03-10T10:11:52.129 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../051-python3-pastedeploy_2.1.1-1_all.deb ... 2026-03-10T10:11:52.130 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-pastedeploy (2.1.1-1) ... 2026-03-10T10:11:52.146 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-webtest. 2026-03-10T10:11:52.151 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../052-python3-webtest_2.0.35-1_all.deb ... 2026-03-10T10:11:52.152 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-webtest (2.0.35-1) ... 2026-03-10T10:11:52.168 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-pecan. 2026-03-10T10:11:52.174 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../053-python3-pecan_1.3.3-4ubuntu2_all.deb ... 2026-03-10T10:11:52.175 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-pecan (1.3.3-4ubuntu2) ... 2026-03-10T10:11:52.208 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-werkzeug. 2026-03-10T10:11:52.214 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../054-python3-werkzeug_2.0.2+dfsg1-1ubuntu0.22.04.3_all.deb ... 2026-03-10T10:11:52.214 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-10T10:11:52.238 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package ceph-mgr-modules-core. 2026-03-10T10:11:52.243 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../055-ceph-mgr-modules-core_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:52.244 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:52.290 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package libsqlite3-mod-ceph. 2026-03-10T10:11:52.293 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../056-libsqlite3-mod-ceph_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:52.374 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:52.493 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package ceph-mgr. 2026-03-10T10:11:52.499 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../057-ceph-mgr_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:52.505 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:52.538 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package ceph-mon. 2026-03-10T10:11:52.544 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../058-ceph-mon_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:52.545 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:52.660 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package libfuse2:amd64. 2026-03-10T10:11:52.666 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../059-libfuse2_2.9.9-5ubuntu3_amd64.deb ... 2026-03-10T10:11:52.667 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-10T10:11:52.688 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package ceph-osd. 2026-03-10T10:11:52.694 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../060-ceph-osd_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:52.695 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:52.847 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:11:52.851 DEBUG:teuthology.parallel:result is None 2026-03-10T10:11:53.032 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package ceph. 2026-03-10T10:11:53.037 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../061-ceph_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:53.039 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:53.058 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package ceph-fuse. 2026-03-10T10:11:53.061 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../062-ceph-fuse_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:53.062 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:53.094 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package ceph-mds. 2026-03-10T10:11:53.098 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../063-ceph-mds_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:53.099 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:53.160 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package cephadm. 2026-03-10T10:11:53.167 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../064-cephadm_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:53.168 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:53.186 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-asyncssh. 2026-03-10T10:11:53.191 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../065-python3-asyncssh_2.5.0-1ubuntu0.1_all.deb ... 2026-03-10T10:11:53.192 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-10T10:11:53.219 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package ceph-mgr-cephadm. 2026-03-10T10:11:53.225 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../066-ceph-mgr-cephadm_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:53.226 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:53.250 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-repoze.lru. 2026-03-10T10:11:53.256 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../067-python3-repoze.lru_0.7-2_all.deb ... 2026-03-10T10:11:53.257 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-repoze.lru (0.7-2) ... 2026-03-10T10:11:53.275 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-routes. 2026-03-10T10:11:53.280 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../068-python3-routes_2.5.1-1ubuntu1_all.deb ... 2026-03-10T10:11:53.281 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-routes (2.5.1-1ubuntu1) ... 2026-03-10T10:11:53.305 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package ceph-mgr-dashboard. 2026-03-10T10:11:53.311 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../069-ceph-mgr-dashboard_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:53.312 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:53.720 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-sklearn-lib:amd64. 2026-03-10T10:11:53.726 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../070-python3-sklearn-lib_0.23.2-5ubuntu6_amd64.deb ... 2026-03-10T10:11:53.727 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-10T10:11:53.795 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-joblib. 2026-03-10T10:11:53.803 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../071-python3-joblib_0.17.0-4ubuntu1_all.deb ... 2026-03-10T10:11:53.804 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-joblib (0.17.0-4ubuntu1) ... 2026-03-10T10:11:53.844 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-threadpoolctl. 2026-03-10T10:11:53.850 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../072-python3-threadpoolctl_3.1.0-1_all.deb ... 2026-03-10T10:11:53.851 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-threadpoolctl (3.1.0-1) ... 2026-03-10T10:11:53.868 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-sklearn. 2026-03-10T10:11:53.875 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../073-python3-sklearn_0.23.2-5ubuntu6_all.deb ... 2026-03-10T10:11:53.875 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-10T10:11:54.026 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package ceph-mgr-diskprediction-local. 2026-03-10T10:11:54.029 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../074-ceph-mgr-diskprediction-local_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:54.030 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:54.314 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-cachetools. 2026-03-10T10:11:54.317 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../075-python3-cachetools_5.0.0-1_all.deb ... 2026-03-10T10:11:54.318 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-cachetools (5.0.0-1) ... 2026-03-10T10:11:54.333 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-rsa. 2026-03-10T10:11:54.339 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../076-python3-rsa_4.8-1_all.deb ... 2026-03-10T10:11:54.340 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-rsa (4.8-1) ... 2026-03-10T10:11:54.359 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-google-auth. 2026-03-10T10:11:54.365 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../077-python3-google-auth_1.5.1-3_all.deb ... 2026-03-10T10:11:54.366 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-google-auth (1.5.1-3) ... 2026-03-10T10:11:54.387 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-requests-oauthlib. 2026-03-10T10:11:54.393 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../078-python3-requests-oauthlib_1.3.0+ds-0.1_all.deb ... 2026-03-10T10:11:54.394 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-10T10:11:54.410 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-websocket. 2026-03-10T10:11:54.416 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../079-python3-websocket_1.2.3-1_all.deb ... 2026-03-10T10:11:54.417 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-websocket (1.2.3-1) ... 2026-03-10T10:11:54.437 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-kubernetes. 2026-03-10T10:11:54.443 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../080-python3-kubernetes_12.0.1-1ubuntu1_all.deb ... 2026-03-10T10:11:54.457 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-10T10:11:54.614 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package ceph-mgr-k8sevents. 2026-03-10T10:11:54.619 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../081-ceph-mgr-k8sevents_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:54.620 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:54.635 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package libonig5:amd64. 2026-03-10T10:11:54.641 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../082-libonig5_6.9.7.1-2build1_amd64.deb ... 2026-03-10T10:11:54.642 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-10T10:11:54.660 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package libjq1:amd64. 2026-03-10T10:11:54.665 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../083-libjq1_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-10T10:11:54.671 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-10T10:11:54.688 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package jq. 2026-03-10T10:11:54.694 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../084-jq_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-10T10:11:54.699 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking jq (1.6-2.1ubuntu3.1) ... 2026-03-10T10:11:54.715 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package socat. 2026-03-10T10:11:54.720 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../085-socat_1.7.4.1-3ubuntu4_amd64.deb ... 2026-03-10T10:11:54.721 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking socat (1.7.4.1-3ubuntu4) ... 2026-03-10T10:11:54.748 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package xmlstarlet. 2026-03-10T10:11:54.754 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../086-xmlstarlet_1.6.1-2.1_amd64.deb ... 2026-03-10T10:11:54.755 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking xmlstarlet (1.6.1-2.1) ... 2026-03-10T10:11:54.801 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package ceph-test. 2026-03-10T10:11:54.806 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../087-ceph-test_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:54.807 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:55.513 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package ceph-volume. 2026-03-10T10:11:55.518 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../088-ceph-volume_19.2.3-678-ge911bdeb-1jammy_all.deb ... 2026-03-10T10:11:55.519 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:55.549 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package libcephfs-dev. 2026-03-10T10:11:55.554 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../089-libcephfs-dev_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:55.556 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:55.575 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package lua-socket:amd64. 2026-03-10T10:11:55.581 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../090-lua-socket_3.0~rc1+git+ac3201d-6_amd64.deb ... 2026-03-10T10:11:55.582 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-10T10:11:55.610 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package lua-sec:amd64. 2026-03-10T10:11:55.616 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../091-lua-sec_1.0.2-1_amd64.deb ... 2026-03-10T10:11:55.617 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking lua-sec:amd64 (1.0.2-1) ... 2026-03-10T10:11:55.638 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package nvme-cli. 2026-03-10T10:11:55.644 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../092-nvme-cli_1.16-3ubuntu0.3_amd64.deb ... 2026-03-10T10:11:55.645 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking nvme-cli (1.16-3ubuntu0.3) ... 2026-03-10T10:11:55.685 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package pkg-config. 2026-03-10T10:11:55.691 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../093-pkg-config_0.29.2-1ubuntu3_amd64.deb ... 2026-03-10T10:11:55.692 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking pkg-config (0.29.2-1ubuntu3) ... 2026-03-10T10:11:55.707 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python-asyncssh-doc. 2026-03-10T10:11:55.713 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../094-python-asyncssh-doc_2.5.0-1ubuntu0.1_all.deb ... 2026-03-10T10:11:55.713 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-10T10:11:55.757 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-iniconfig. 2026-03-10T10:11:55.762 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../095-python3-iniconfig_1.1.1-2_all.deb ... 2026-03-10T10:11:55.763 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-iniconfig (1.1.1-2) ... 2026-03-10T10:11:55.778 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-pastescript. 2026-03-10T10:11:55.783 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../096-python3-pastescript_2.0.2-4_all.deb ... 2026-03-10T10:11:55.784 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-pastescript (2.0.2-4) ... 2026-03-10T10:11:55.804 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-pluggy. 2026-03-10T10:11:55.809 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../097-python3-pluggy_0.13.0-7.1_all.deb ... 2026-03-10T10:11:55.811 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-pluggy (0.13.0-7.1) ... 2026-03-10T10:11:55.829 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-psutil. 2026-03-10T10:11:55.835 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../098-python3-psutil_5.9.0-1build1_amd64.deb ... 2026-03-10T10:11:55.836 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-psutil (5.9.0-1build1) ... 2026-03-10T10:11:55.859 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-py. 2026-03-10T10:11:55.864 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../099-python3-py_1.10.0-1_all.deb ... 2026-03-10T10:11:55.865 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-py (1.10.0-1) ... 2026-03-10T10:11:55.889 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-pygments. 2026-03-10T10:11:55.894 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../100-python3-pygments_2.11.2+dfsg-2ubuntu0.1_all.deb ... 2026-03-10T10:11:55.895 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-10T10:11:55.956 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-pyinotify. 2026-03-10T10:11:55.961 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../101-python3-pyinotify_0.9.6-1.3_all.deb ... 2026-03-10T10:11:55.963 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-pyinotify (0.9.6-1.3) ... 2026-03-10T10:11:55.977 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-toml. 2026-03-10T10:11:55.983 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../102-python3-toml_0.10.2-1_all.deb ... 2026-03-10T10:11:55.984 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-toml (0.10.2-1) ... 2026-03-10T10:11:56.000 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-pytest. 2026-03-10T10:11:56.005 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../103-python3-pytest_6.2.5-1ubuntu2_all.deb ... 2026-03-10T10:11:56.006 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-pytest (6.2.5-1ubuntu2) ... 2026-03-10T10:11:56.033 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-simplejson. 2026-03-10T10:11:56.038 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../104-python3-simplejson_3.17.6-1build1_amd64.deb ... 2026-03-10T10:11:56.039 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-simplejson (3.17.6-1build1) ... 2026-03-10T10:11:56.060 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package qttranslations5-l10n. 2026-03-10T10:11:56.066 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../105-qttranslations5-l10n_5.15.3-1_all.deb ... 2026-03-10T10:11:56.067 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking qttranslations5-l10n (5.15.3-1) ... 2026-03-10T10:11:56.182 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package radosgw. 2026-03-10T10:11:56.185 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../106-radosgw_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:56.186 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:56.391 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package rbd-fuse. 2026-03-10T10:11:56.396 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../107-rbd-fuse_19.2.3-678-ge911bdeb-1jammy_amd64.deb ... 2026-03-10T10:11:56.397 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:56.415 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package smartmontools. 2026-03-10T10:11:56.421 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../108-smartmontools_7.2-1ubuntu0.1_amd64.deb ... 2026-03-10T10:11:56.430 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking smartmontools (7.2-1ubuntu0.1) ... 2026-03-10T10:11:56.473 INFO:teuthology.orchestra.run.vm03.stdout:Setting up smartmontools (7.2-1ubuntu0.1) ... 2026-03-10T10:11:56.724 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/smartd.service → /lib/systemd/system/smartmontools.service. 2026-03-10T10:11:56.724 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartmontools.service → /lib/systemd/system/smartmontools.service. 2026-03-10T10:11:57.097 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-iniconfig (1.1.1-2) ... 2026-03-10T10:11:57.159 INFO:teuthology.orchestra.run.vm03.stdout:Setting up libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-10T10:11:57.161 INFO:teuthology.orchestra.run.vm03.stdout:Setting up nvme-cli (1.16-3ubuntu0.3) ... 2026-03-10T10:11:57.221 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /lib/systemd/system/nvmefc-boot-connections.service. 2026-03-10T10:11:57.438 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmf-autoconnect.service → /lib/systemd/system/nvmf-autoconnect.service. 2026-03-10T10:11:57.825 INFO:teuthology.orchestra.run.vm03.stdout:nvmf-connect.target is a disabled or a static unit, not starting it. 2026-03-10T10:11:57.831 INFO:teuthology.orchestra.run.vm03.stdout:Could not execute systemctl: at /usr/bin/deb-systemd-invoke line 142. 2026-03-10T10:11:57.833 INFO:teuthology.orchestra.run.vm03.stdout:Setting up cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:57.871 INFO:teuthology.orchestra.run.vm03.stdout:Adding system user cephadm....done 2026-03-10T10:11:57.879 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-10T10:11:57.952 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-jaraco.classes (3.2.1-3) ... 2026-03-10T10:11:58.031 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-10T10:11:58.039 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-jaraco.functools (3.4.0-2) ... 2026-03-10T10:11:58.107 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-repoze.lru (0.7-2) ... 2026-03-10T10:11:58.177 INFO:teuthology.orchestra.run.vm03.stdout:Setting up liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-10T10:11:58.180 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-py (1.10.0-1) ... 2026-03-10T10:11:58.275 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-joblib (0.17.0-4ubuntu1) ... 2026-03-10T10:11:58.392 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-cachetools (5.0.0-1) ... 2026-03-10T10:11:58.460 INFO:teuthology.orchestra.run.vm03.stdout:Setting up unzip (6.0-26ubuntu3.2) ... 2026-03-10T10:11:58.468 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-pyinotify (0.9.6-1.3) ... 2026-03-10T10:11:58.538 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-threadpoolctl (3.1.0-1) ... 2026-03-10T10:11:58.617 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:11:58.696 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-10T10:11:58.698 INFO:teuthology.orchestra.run.vm03.stdout:Setting up libnbd0 (1.10.5-1) ... 2026-03-10T10:11:58.701 INFO:teuthology.orchestra.run.vm03.stdout:Setting up lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-10T10:11:58.704 INFO:teuthology.orchestra.run.vm03.stdout:Setting up libreadline-dev:amd64 (8.1.2-1) ... 2026-03-10T10:11:58.706 INFO:teuthology.orchestra.run.vm03.stdout:Setting up libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-10T10:11:58.708 INFO:teuthology.orchestra.run.vm03.stdout:Setting up lua5.1 (5.1.5-8.1build4) ... 2026-03-10T10:11:58.713 INFO:teuthology.orchestra.run.vm03.stdout:update-alternatives: using /usr/bin/lua5.1 to provide /usr/bin/lua (lua-interpreter) in auto mode 2026-03-10T10:11:58.716 INFO:teuthology.orchestra.run.vm03.stdout:update-alternatives: using /usr/bin/luac5.1 to provide /usr/bin/luac (lua-compiler) in auto mode 2026-03-10T10:11:58.719 INFO:teuthology.orchestra.run.vm03.stdout:Setting up libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-10T10:11:58.722 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-psutil (5.9.0-1build1) ... 2026-03-10T10:11:58.858 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-natsort (8.0.2-1) ... 2026-03-10T10:11:58.933 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-routes (2.5.1-1ubuntu1) ... 2026-03-10T10:11:59.005 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-simplejson (3.17.6-1build1) ... 2026-03-10T10:11:59.086 INFO:teuthology.orchestra.run.vm03.stdout:Setting up zip (3.0-12build2) ... 2026-03-10T10:11:59.089 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-10T10:11:59.365 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-tempita (0.5.2-6ubuntu1) ... 2026-03-10T10:11:59.438 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python-pastedeploy-tpl (2.1.1-1) ... 2026-03-10T10:11:59.440 INFO:teuthology.orchestra.run.vm03.stdout:Setting up qttranslations5-l10n (5.15.3-1) ... 2026-03-10T10:11:59.442 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-10T10:11:59.532 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-10T10:11:59.665 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-paste (3.5.0+dfsg1-1) ... 2026-03-10T10:11:59.795 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-10T10:11:59.883 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-10T10:12:00.010 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-jaraco.text (3.6.0-2) ... 2026-03-10T10:12:00.068 INFO:teuthology.orchestra.run.vm03.stdout:Setting up socat (1.7.4.1-3ubuntu4) ... 2026-03-10T10:12:00.071 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:00.162 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-10T10:12:00.730 INFO:teuthology.orchestra.run.vm03.stdout:Setting up pkg-config (0.29.2-1ubuntu3) ... 2026-03-10T10:12:00.752 INFO:teuthology.orchestra.run.vm03.stdout:Setting up libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:12:00.756 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-toml (0.10.2-1) ... 2026-03-10T10:12:00.827 INFO:teuthology.orchestra.run.vm03.stdout:Setting up librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-10T10:12:00.829 INFO:teuthology.orchestra.run.vm03.stdout:Setting up xmlstarlet (1.6.1-2.1) ... 2026-03-10T10:12:00.831 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-pluggy (0.13.0-7.1) ... 2026-03-10T10:12:00.901 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-zc.lockfile (2.0-1) ... 2026-03-10T10:12:00.967 INFO:teuthology.orchestra.run.vm03.stdout:Setting up libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:12:00.970 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-rsa (4.8-1) ... 2026-03-10T10:12:01.041 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-singledispatch (3.4.0.3-3) ... 2026-03-10T10:12:01.108 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-logutils (0.3.3-8) ... 2026-03-10T10:12:01.188 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-tempora (4.1.2-1) ... 2026-03-10T10:12:01.254 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-simplegeneric (0.8.1-3) ... 2026-03-10T10:12:01.319 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-prettytable (2.5.0-2) ... 2026-03-10T10:12:01.390 INFO:teuthology.orchestra.run.vm03.stdout:Setting up liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-10T10:12:01.392 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-websocket (1.2.3-1) ... 2026-03-10T10:12:01.472 INFO:teuthology.orchestra.run.vm03.stdout:Setting up libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-10T10:12:01.474 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-10T10:12:01.550 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-10T10:12:01.641 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-10T10:12:01.739 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-jaraco.collections (3.4.0-2) ... 2026-03-10T10:12:01.807 INFO:teuthology.orchestra.run.vm03.stdout:Setting up liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-10T10:12:01.809 INFO:teuthology.orchestra.run.vm03.stdout:Setting up lua-sec:amd64 (1.0.2-1) ... 2026-03-10T10:12:01.811 INFO:teuthology.orchestra.run.vm03.stdout:Setting up libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-10T10:12:01.814 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-pytest (6.2.5-1ubuntu2) ... 2026-03-10T10:12:01.955 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-pastedeploy (2.1.1-1) ... 2026-03-10T10:12:02.026 INFO:teuthology.orchestra.run.vm03.stdout:Setting up lua-any (27ubuntu1) ... 2026-03-10T10:12:02.028 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-portend (3.0.0-1) ... 2026-03-10T10:12:02.097 INFO:teuthology.orchestra.run.vm03.stdout:Setting up libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:12:02.099 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-google-auth (1.5.1-3) ... 2026-03-10T10:12:02.182 INFO:teuthology.orchestra.run.vm03.stdout:Setting up jq (1.6-2.1ubuntu3.1) ... 2026-03-10T10:12:02.184 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-webtest (2.0.35-1) ... 2026-03-10T10:12:02.309 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-cherrypy3 (18.6.1-4) ... 2026-03-10T10:12:02.474 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-pastescript (2.0.2-4) ... 2026-03-10T10:12:02.562 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-pecan (1.3.3-4ubuntu2) ... 2026-03-10T10:12:02.680 INFO:teuthology.orchestra.run.vm03.stdout:Setting up libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-10T10:12:02.682 INFO:teuthology.orchestra.run.vm03.stdout:Setting up librados2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:02.685 INFO:teuthology.orchestra.run.vm03.stdout:Setting up libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:02.687 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-10T10:12:03.293 INFO:teuthology.orchestra.run.vm03.stdout:Setting up luarocks (3.8.0+dfsg1-1) ... 2026-03-10T10:12:03.301 INFO:teuthology.orchestra.run.vm03.stdout:Setting up libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:03.303 INFO:teuthology.orchestra.run.vm03.stdout:Setting up libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:03.305 INFO:teuthology.orchestra.run.vm03.stdout:Setting up librbd1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:03.307 INFO:teuthology.orchestra.run.vm03.stdout:Setting up ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:03.310 INFO:teuthology.orchestra.run.vm03.stdout:Setting up ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:03.374 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/remote-fs.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-10T10:12:03.374 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-10T10:12:03.769 INFO:teuthology.orchestra.run.vm03.stdout:Setting up libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:03.771 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:03.773 INFO:teuthology.orchestra.run.vm03.stdout:Setting up librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:03.775 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:03.778 INFO:teuthology.orchestra.run.vm03.stdout:Setting up rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:03.780 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:03.782 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:03.784 INFO:teuthology.orchestra.run.vm03.stdout:Setting up ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:03.818 INFO:teuthology.orchestra.run.vm03.stdout:Adding group ceph....done 2026-03-10T10:12:03.855 INFO:teuthology.orchestra.run.vm03.stdout:Adding system user ceph....done 2026-03-10T10:12:03.863 INFO:teuthology.orchestra.run.vm03.stdout:Setting system user ceph properties....done 2026-03-10T10:12:03.867 INFO:teuthology.orchestra.run.vm03.stdout:chown: cannot access '/var/log/ceph/*.log*': No such file or directory 2026-03-10T10:12:03.931 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /lib/systemd/system/ceph.target. 2026-03-10T10:12:04.149 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/rbdmap.service → /lib/systemd/system/rbdmap.service. 2026-03-10T10:12:04.535 INFO:teuthology.orchestra.run.vm03.stdout:Setting up ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:04.537 INFO:teuthology.orchestra.run.vm03.stdout:Setting up radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:04.763 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-10T10:12:04.763 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-10T10:12:05.123 INFO:teuthology.orchestra.run.vm03.stdout:Setting up ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:05.207 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /lib/systemd/system/ceph-crash.service. 2026-03-10T10:12:05.584 INFO:teuthology.orchestra.run.vm03.stdout:Setting up ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:05.643 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-10T10:12:05.643 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-10T10:12:06.010 INFO:teuthology.orchestra.run.vm03.stdout:Setting up ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:06.078 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-10T10:12:06.078 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-10T10:12:06.523 INFO:teuthology.orchestra.run.vm03.stdout:Setting up ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:06.613 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-10T10:12:06.614 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-10T10:12:07.208 INFO:teuthology.orchestra.run.vm03.stdout:Setting up ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:07.284 INFO:teuthology.orchestra.run.vm03.stdout:Setting up ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:07.346 INFO:teuthology.orchestra.run.vm03.stdout:Setting up ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:07.413 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-10T10:12:07.413 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-10T10:12:07.943 INFO:teuthology.orchestra.run.vm03.stdout:Setting up ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:08.017 INFO:teuthology.orchestra.run.vm03.stdout:Setting up ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:08.021 INFO:teuthology.orchestra.run.vm03.stdout:Setting up ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:08.062 INFO:teuthology.orchestra.run.vm03.stdout:Setting up ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:12:08.186 INFO:teuthology.orchestra.run.vm03.stdout:Processing triggers for mailcap (3.70+nmu1ubuntu1) ... 2026-03-10T10:12:08.201 INFO:teuthology.orchestra.run.vm03.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T10:12:08.226 INFO:teuthology.orchestra.run.vm03.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:12:08.385 INFO:teuthology.orchestra.run.vm03.stdout:Processing triggers for install-info (6.8-4build1) ... 2026-03-10T10:12:08.709 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-10T10:12:08.709 INFO:teuthology.orchestra.run.vm03.stdout:Running kernel seems to be up-to-date. 2026-03-10T10:12:08.709 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-10T10:12:08.709 INFO:teuthology.orchestra.run.vm03.stdout:Services to be restarted: 2026-03-10T10:12:08.715 INFO:teuthology.orchestra.run.vm03.stdout: systemctl restart packagekit.service 2026-03-10T10:12:08.717 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-10T10:12:08.718 INFO:teuthology.orchestra.run.vm03.stdout:Service restarts being deferred: 2026-03-10T10:12:08.718 INFO:teuthology.orchestra.run.vm03.stdout: systemctl restart networkd-dispatcher.service 2026-03-10T10:12:08.718 INFO:teuthology.orchestra.run.vm03.stdout: systemctl restart unattended-upgrades.service 2026-03-10T10:12:08.718 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-10T10:12:08.718 INFO:teuthology.orchestra.run.vm03.stdout:No containers need to be restarted. 2026-03-10T10:12:08.718 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-10T10:12:08.718 INFO:teuthology.orchestra.run.vm03.stdout:No user sessions are running outdated binaries. 2026-03-10T10:12:08.718 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-10T10:12:08.718 INFO:teuthology.orchestra.run.vm03.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-10T10:12:09.696 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:12:09.699 DEBUG:teuthology.orchestra.run.vm03:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install python3-xmltodict python3-jmespath 2026-03-10T10:12:09.779 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:12:10.014 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:12:10.014 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:12:10.296 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:12:10.296 INFO:teuthology.orchestra.run.vm03.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T10:12:10.297 INFO:teuthology.orchestra.run.vm03.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-10T10:12:10.297 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:12:10.316 INFO:teuthology.orchestra.run.vm03.stdout:The following NEW packages will be installed: 2026-03-10T10:12:10.317 INFO:teuthology.orchestra.run.vm03.stdout: python3-jmespath python3-xmltodict 2026-03-10T10:12:10.775 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 2 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:12:10.775 INFO:teuthology.orchestra.run.vm03.stdout:Need to get 34.3 kB of archives. 2026-03-10T10:12:10.775 INFO:teuthology.orchestra.run.vm03.stdout:After this operation, 146 kB of additional disk space will be used. 2026-03-10T10:12:10.775 INFO:teuthology.orchestra.run.vm03.stdout:Get:1 https://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jmespath all 0.10.0-1 [21.7 kB] 2026-03-10T10:12:10.856 INFO:teuthology.orchestra.run.vm03.stdout:Get:2 https://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-xmltodict all 0.12.0-2 [12.6 kB] 2026-03-10T10:12:11.076 INFO:teuthology.orchestra.run.vm03.stdout:Fetched 34.3 kB in 1s (64.2 kB/s) 2026-03-10T10:12:11.092 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-jmespath. 2026-03-10T10:12:11.123 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118577 files and directories currently installed.) 2026-03-10T10:12:11.126 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../python3-jmespath_0.10.0-1_all.deb ... 2026-03-10T10:12:11.127 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-jmespath (0.10.0-1) ... 2026-03-10T10:12:11.144 INFO:teuthology.orchestra.run.vm03.stdout:Selecting previously unselected package python3-xmltodict. 2026-03-10T10:12:11.150 INFO:teuthology.orchestra.run.vm03.stdout:Preparing to unpack .../python3-xmltodict_0.12.0-2_all.deb ... 2026-03-10T10:12:11.150 INFO:teuthology.orchestra.run.vm03.stdout:Unpacking python3-xmltodict (0.12.0-2) ... 2026-03-10T10:12:11.177 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-xmltodict (0.12.0-2) ... 2026-03-10T10:12:11.242 INFO:teuthology.orchestra.run.vm03.stdout:Setting up python3-jmespath (0.10.0-1) ... 2026-03-10T10:12:11.576 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-10T10:12:11.576 INFO:teuthology.orchestra.run.vm03.stdout:Running kernel seems to be up-to-date. 2026-03-10T10:12:11.576 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-10T10:12:11.576 INFO:teuthology.orchestra.run.vm03.stdout:Services to be restarted: 2026-03-10T10:12:11.583 INFO:teuthology.orchestra.run.vm03.stdout: systemctl restart packagekit.service 2026-03-10T10:12:11.587 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-10T10:12:11.587 INFO:teuthology.orchestra.run.vm03.stdout:Service restarts being deferred: 2026-03-10T10:12:11.587 INFO:teuthology.orchestra.run.vm03.stdout: systemctl restart networkd-dispatcher.service 2026-03-10T10:12:11.587 INFO:teuthology.orchestra.run.vm03.stdout: systemctl restart unattended-upgrades.service 2026-03-10T10:12:11.587 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-10T10:12:11.587 INFO:teuthology.orchestra.run.vm03.stdout:No containers need to be restarted. 2026-03-10T10:12:11.587 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-10T10:12:11.587 INFO:teuthology.orchestra.run.vm03.stdout:No user sessions are running outdated binaries. 2026-03-10T10:12:11.587 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-10T10:12:11.587 INFO:teuthology.orchestra.run.vm03.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-10T10:12:12.590 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:12:12.594 DEBUG:teuthology.parallel:result is None 2026-03-10T10:12:12.594 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T10:12:13.189 DEBUG:teuthology.orchestra.run.vm00:> dpkg-query -W -f '${Version}' ceph 2026-03-10T10:12:13.197 INFO:teuthology.orchestra.run.vm00.stdout:19.2.3-678-ge911bdeb-1jammy 2026-03-10T10:12:13.198 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678-ge911bdeb-1jammy 2026-03-10T10:12:13.198 INFO:teuthology.task.install:The correct ceph version 19.2.3-678-ge911bdeb-1jammy is installed. 2026-03-10T10:12:13.199 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T10:12:13.824 DEBUG:teuthology.orchestra.run.vm03:> dpkg-query -W -f '${Version}' ceph 2026-03-10T10:12:13.835 INFO:teuthology.orchestra.run.vm03.stdout:19.2.3-678-ge911bdeb-1jammy 2026-03-10T10:12:13.835 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678-ge911bdeb-1jammy 2026-03-10T10:12:13.835 INFO:teuthology.task.install:The correct ceph version 19.2.3-678-ge911bdeb-1jammy is installed. 2026-03-10T10:12:13.836 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T10:12:14.473 DEBUG:teuthology.orchestra.run.vm09:> dpkg-query -W -f '${Version}' ceph 2026-03-10T10:12:14.482 INFO:teuthology.orchestra.run.vm09.stdout:19.2.3-678-ge911bdeb-1jammy 2026-03-10T10:12:14.482 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678-ge911bdeb-1jammy 2026-03-10T10:12:14.482 INFO:teuthology.task.install:The correct ceph version 19.2.3-678-ge911bdeb-1jammy is installed. 2026-03-10T10:12:14.483 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-03-10T10:12:14.483 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-10T10:12:14.483 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-10T10:12:14.491 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-10T10:12:14.491 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-10T10:12:14.499 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T10:12:14.499 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-10T10:12:14.534 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-03-10T10:12:14.534 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-10T10:12:14.534 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/usr/bin/daemon-helper 2026-03-10T10:12:14.543 DEBUG:teuthology.orchestra.run.vm00:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-10T10:12:14.594 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-10T10:12:14.595 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/usr/bin/daemon-helper 2026-03-10T10:12:14.602 DEBUG:teuthology.orchestra.run.vm03:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-10T10:12:14.654 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T10:12:14.654 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/usr/bin/daemon-helper 2026-03-10T10:12:14.662 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-10T10:12:14.710 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-03-10T10:12:14.710 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-10T10:12:14.711 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-10T10:12:14.718 DEBUG:teuthology.orchestra.run.vm00:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-10T10:12:14.766 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-10T10:12:14.766 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-10T10:12:14.774 DEBUG:teuthology.orchestra.run.vm03:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-10T10:12:14.826 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T10:12:14.826 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-10T10:12:14.834 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-10T10:12:14.886 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-03-10T10:12:14.886 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-10T10:12:14.886 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/usr/bin/stdin-killer 2026-03-10T10:12:14.894 DEBUG:teuthology.orchestra.run.vm00:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-10T10:12:14.942 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-10T10:12:14.942 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/usr/bin/stdin-killer 2026-03-10T10:12:14.950 DEBUG:teuthology.orchestra.run.vm03:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-10T10:12:15.001 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T10:12:15.001 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/usr/bin/stdin-killer 2026-03-10T10:12:15.010 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-10T10:12:15.058 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-10T10:12:15.102 INFO:tasks.cephadm:Config: {'conf': {'global': {'mon election default strategy': 3}, 'mgr': {'debug mgr': 20, 'debug ms': 1, 'mgr/cephadm/use_agent': True}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'MON_DOWN', 'POOL_APP_NOT_ENABLED', 'mon down', 'mons down', 'out of quorum', 'CEPHADM_FAILED_DAEMON'], 'log-only-match': ['CEPHADM_'], 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-10T10:12:15.102 INFO:tasks.cephadm:Cluster image is quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T10:12:15.102 INFO:tasks.cephadm:Cluster fsid is 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:12:15.102 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-10T10:12:15.102 INFO:tasks.cephadm:Monitor IPs: {'mon.a': '192.168.123.100', 'mon.b': '192.168.123.103', 'mon.c': '192.168.123.109'} 2026-03-10T10:12:15.102 INFO:tasks.cephadm:First mon is mon.a on vm00 2026-03-10T10:12:15.102 INFO:tasks.cephadm:First mgr is a 2026-03-10T10:12:15.102 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-10T10:12:15.103 DEBUG:teuthology.orchestra.run.vm00:> sudo hostname $(hostname -s) 2026-03-10T10:12:15.111 DEBUG:teuthology.orchestra.run.vm03:> sudo hostname $(hostname -s) 2026-03-10T10:12:15.120 DEBUG:teuthology.orchestra.run.vm09:> sudo hostname $(hostname -s) 2026-03-10T10:12:15.130 INFO:tasks.cephadm:Downloading "compiled" cephadm from cachra 2026-03-10T10:12:15.130 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T10:12:15.745 INFO:tasks.cephadm:builder_project result: [{'url': 'https://1.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default/', 'chacra_url': 'https://1.chacra.ceph.com/repos/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/flavors/default/', 'ref': 'squid', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'distro': 'ubuntu', 'distro_version': '22.04', 'distro_codename': 'jammy', 'modified': '2026-02-25 19:37:07.680480', 'status': 'ready', 'flavor': 'default', 'project': 'ceph', 'archs': ['x86_64'], 'extra': {'version': '19.2.3-678-ge911bdeb', 'package_manager_version': '19.2.3-678-ge911bdeb-1jammy', 'build_url': 'https://jenkins.ceph.com/job/ceph-dev-pipeline/3275/', 'root_build_cause': '', 'node_name': '10.20.192.98+toko08', 'job_name': 'ceph-dev-pipeline'}}] 2026-03-10T10:12:16.413 INFO:tasks.util.chacra:got chacra host 1.chacra.ceph.com, ref squid, sha1 e911bdebe5c8faa3800735d1568fcdca65db60df from https://shaman.ceph.com/api/search/?project=ceph&distros=ubuntu%2F22.04%2Fx86_64&flavor=default&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T10:12:16.414 INFO:tasks.cephadm:Discovered cachra url: https://1.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/x86_64/flavors/default/cephadm 2026-03-10T10:12:16.414 INFO:tasks.cephadm:Downloading cephadm from url: https://1.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/x86_64/flavors/default/cephadm 2026-03-10T10:12:16.414 DEBUG:teuthology.orchestra.run.vm00:> curl --silent -L https://1.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-10T10:12:19.692 INFO:teuthology.orchestra.run.vm00.stdout:-rw-rw-r-- 1 ubuntu ubuntu 795696 Mar 10 10:12 /home/ubuntu/cephtest/cephadm 2026-03-10T10:12:19.692 DEBUG:teuthology.orchestra.run.vm03:> curl --silent -L https://1.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-10T10:12:20.994 INFO:teuthology.orchestra.run.vm03.stdout:-rw-rw-r-- 1 ubuntu ubuntu 795696 Mar 10 10:12 /home/ubuntu/cephtest/cephadm 2026-03-10T10:12:20.994 DEBUG:teuthology.orchestra.run.vm09:> curl --silent -L https://1.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/ubuntu/jammy/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-10T10:12:23.558 INFO:teuthology.orchestra.run.vm09.stdout:-rw-rw-r-- 1 ubuntu ubuntu 795696 Mar 10 10:12 /home/ubuntu/cephtest/cephadm 2026-03-10T10:12:23.558 DEBUG:teuthology.orchestra.run.vm00:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-10T10:12:23.563 DEBUG:teuthology.orchestra.run.vm03:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-10T10:12:23.567 DEBUG:teuthology.orchestra.run.vm09:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-10T10:12:23.576 INFO:tasks.cephadm:Pulling image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df on all hosts... 2026-03-10T10:12:23.576 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-10T10:12:23.608 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-10T10:12:23.610 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-10T10:12:23.701 INFO:teuthology.orchestra.run.vm09.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T10:12:23.706 INFO:teuthology.orchestra.run.vm00.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T10:12:23.709 INFO:teuthology.orchestra.run.vm03.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T10:13:33.798 INFO:teuthology.orchestra.run.vm09.stdout:{ 2026-03-10T10:13:33.799 INFO:teuthology.orchestra.run.vm09.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-10T10:13:33.799 INFO:teuthology.orchestra.run.vm09.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-10T10:13:33.799 INFO:teuthology.orchestra.run.vm09.stdout: "repo_digests": [ 2026-03-10T10:13:33.799 INFO:teuthology.orchestra.run.vm09.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-10T10:13:33.799 INFO:teuthology.orchestra.run.vm09.stdout: ] 2026-03-10T10:13:33.799 INFO:teuthology.orchestra.run.vm09.stdout:} 2026-03-10T10:13:46.968 INFO:teuthology.orchestra.run.vm00.stdout:{ 2026-03-10T10:13:46.968 INFO:teuthology.orchestra.run.vm00.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-10T10:13:46.968 INFO:teuthology.orchestra.run.vm00.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-10T10:13:46.968 INFO:teuthology.orchestra.run.vm00.stdout: "repo_digests": [ 2026-03-10T10:13:46.968 INFO:teuthology.orchestra.run.vm00.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-10T10:13:46.968 INFO:teuthology.orchestra.run.vm00.stdout: ] 2026-03-10T10:13:46.968 INFO:teuthology.orchestra.run.vm00.stdout:} 2026-03-10T10:14:42.102 INFO:teuthology.orchestra.run.vm03.stdout:{ 2026-03-10T10:14:42.102 INFO:teuthology.orchestra.run.vm03.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-10T10:14:42.102 INFO:teuthology.orchestra.run.vm03.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-10T10:14:42.102 INFO:teuthology.orchestra.run.vm03.stdout: "repo_digests": [ 2026-03-10T10:14:42.102 INFO:teuthology.orchestra.run.vm03.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-10T10:14:42.102 INFO:teuthology.orchestra.run.vm03.stdout: ] 2026-03-10T10:14:42.102 INFO:teuthology.orchestra.run.vm03.stdout:} 2026-03-10T10:14:42.120 DEBUG:teuthology.orchestra.run.vm00:> sudo mkdir -p /etc/ceph 2026-03-10T10:14:42.127 DEBUG:teuthology.orchestra.run.vm03:> sudo mkdir -p /etc/ceph 2026-03-10T10:14:42.135 DEBUG:teuthology.orchestra.run.vm09:> sudo mkdir -p /etc/ceph 2026-03-10T10:14:42.142 DEBUG:teuthology.orchestra.run.vm00:> sudo chmod 777 /etc/ceph 2026-03-10T10:14:42.177 DEBUG:teuthology.orchestra.run.vm03:> sudo chmod 777 /etc/ceph 2026-03-10T10:14:42.184 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod 777 /etc/ceph 2026-03-10T10:14:42.192 INFO:tasks.cephadm:Writing seed config... 2026-03-10T10:14:42.193 INFO:tasks.cephadm: override: [global] mon election default strategy = 3 2026-03-10T10:14:42.193 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-10T10:14:42.193 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-10T10:14:42.193 INFO:tasks.cephadm: override: [mgr] mgr/cephadm/use_agent = True 2026-03-10T10:14:42.193 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-10T10:14:42.193 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-10T10:14:42.193 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-10T10:14:42.193 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-10T10:14:42.193 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-10T10:14:42.193 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-10T10:14:42.193 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-10T10:14:42.193 DEBUG:teuthology.orchestra.run.vm00:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-10T10:14:42.222 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad mon election default strategy = 3 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = true bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 mgr/cephadm/use_agent = True [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-10T10:14:42.222 DEBUG:teuthology.orchestra.run.vm00:mon.a> sudo journalctl -f -n 0 -u ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mon.a.service 2026-03-10T10:14:42.264 DEBUG:teuthology.orchestra.run.vm00:mgr.a> sudo journalctl -f -n 0 -u ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mgr.a.service 2026-03-10T10:14:42.308 INFO:tasks.cephadm:Bootstrapping... 2026-03-10T10:14:42.308 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df -v bootstrap --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-id a --mgr-id a --orphan-initial-daemons --skip-monitoring-stack --mon-ip 192.168.123.100 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:14:42.442 INFO:teuthology.orchestra.run.vm00.stdout:-------------------------------------------------------------------------------- 2026-03-10T10:14:42.442 INFO:teuthology.orchestra.run.vm00.stdout:cephadm ['--image', 'quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df', '-v', 'bootstrap', '--fsid', '9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-id', 'a', '--mgr-id', 'a', '--orphan-initial-daemons', '--skip-monitoring-stack', '--mon-ip', '192.168.123.100', '--skip-admin-label'] 2026-03-10T10:14:42.442 INFO:teuthology.orchestra.run.vm00.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-10T10:14:42.443 INFO:teuthology.orchestra.run.vm00.stdout:Verifying podman|docker is present... 2026-03-10T10:14:42.443 INFO:teuthology.orchestra.run.vm00.stdout:Verifying lvm2 is present... 2026-03-10T10:14:42.443 INFO:teuthology.orchestra.run.vm00.stdout:Verifying time synchronization is in place... 2026-03-10T10:14:42.447 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-10T10:14:42.447 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-10T10:14:42.450 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-10T10:14:42.450 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout inactive 2026-03-10T10:14:42.453 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl is-enabled chronyd.service 2026-03-10T10:14:42.453 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Failed to get unit file state for chronyd.service: No such file or directory 2026-03-10T10:14:42.456 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 3 from systemctl is-active chronyd.service 2026-03-10T10:14:42.456 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout inactive 2026-03-10T10:14:42.459 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl is-enabled systemd-timesyncd.service 2026-03-10T10:14:42.459 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout masked 2026-03-10T10:14:42.462 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 3 from systemctl is-active systemd-timesyncd.service 2026-03-10T10:14:42.462 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout inactive 2026-03-10T10:14:42.465 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl is-enabled ntpd.service 2026-03-10T10:14:42.465 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Failed to get unit file state for ntpd.service: No such file or directory 2026-03-10T10:14:42.468 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 3 from systemctl is-active ntpd.service 2026-03-10T10:14:42.468 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout inactive 2026-03-10T10:14:42.472 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout enabled 2026-03-10T10:14:42.474 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout active 2026-03-10T10:14:42.474 INFO:teuthology.orchestra.run.vm00.stdout:Unit ntp.service is enabled and running 2026-03-10T10:14:42.474 INFO:teuthology.orchestra.run.vm00.stdout:Repeating the final host check... 2026-03-10T10:14:42.474 INFO:teuthology.orchestra.run.vm00.stdout:docker (/usr/bin/docker) is present 2026-03-10T10:14:42.474 INFO:teuthology.orchestra.run.vm00.stdout:systemctl is present 2026-03-10T10:14:42.474 INFO:teuthology.orchestra.run.vm00.stdout:lvcreate is present 2026-03-10T10:14:42.476 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-10T10:14:42.476 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-10T10:14:42.479 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-10T10:14:42.479 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout inactive 2026-03-10T10:14:42.481 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl is-enabled chronyd.service 2026-03-10T10:14:42.481 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Failed to get unit file state for chronyd.service: No such file or directory 2026-03-10T10:14:42.483 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 3 from systemctl is-active chronyd.service 2026-03-10T10:14:42.483 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout inactive 2026-03-10T10:14:42.486 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl is-enabled systemd-timesyncd.service 2026-03-10T10:14:42.486 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout masked 2026-03-10T10:14:42.488 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 3 from systemctl is-active systemd-timesyncd.service 2026-03-10T10:14:42.488 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout inactive 2026-03-10T10:14:42.490 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl is-enabled ntpd.service 2026-03-10T10:14:42.490 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Failed to get unit file state for ntpd.service: No such file or directory 2026-03-10T10:14:42.493 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 3 from systemctl is-active ntpd.service 2026-03-10T10:14:42.493 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout inactive 2026-03-10T10:14:42.496 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout enabled 2026-03-10T10:14:42.499 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout active 2026-03-10T10:14:42.499 INFO:teuthology.orchestra.run.vm00.stdout:Unit ntp.service is enabled and running 2026-03-10T10:14:42.499 INFO:teuthology.orchestra.run.vm00.stdout:Host looks OK 2026-03-10T10:14:42.499 INFO:teuthology.orchestra.run.vm00.stdout:Cluster fsid: 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:14:42.499 INFO:teuthology.orchestra.run.vm00.stdout:Acquiring lock 139762884774448 on /run/cephadm/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad.lock 2026-03-10T10:14:42.499 INFO:teuthology.orchestra.run.vm00.stdout:Lock 139762884774448 acquired on /run/cephadm/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad.lock 2026-03-10T10:14:42.499 INFO:teuthology.orchestra.run.vm00.stdout:Verifying IP 192.168.123.100 port 3300 ... 2026-03-10T10:14:42.499 INFO:teuthology.orchestra.run.vm00.stdout:Verifying IP 192.168.123.100 port 6789 ... 2026-03-10T10:14:42.499 INFO:teuthology.orchestra.run.vm00.stdout:Base mon IP(s) is [192.168.123.100:3300, 192.168.123.100:6789], mon addrv is [v2:192.168.123.100:3300,v1:192.168.123.100:6789] 2026-03-10T10:14:42.500 INFO:teuthology.orchestra.run.vm00.stdout:/usr/sbin/ip: stdout default via 192.168.123.1 dev ens3 proto dhcp src 192.168.123.100 metric 100 2026-03-10T10:14:42.500 INFO:teuthology.orchestra.run.vm00.stdout:/usr/sbin/ip: stdout 172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.0.1 linkdown 2026-03-10T10:14:42.500 INFO:teuthology.orchestra.run.vm00.stdout:/usr/sbin/ip: stdout 192.168.123.0/24 dev ens3 proto kernel scope link src 192.168.123.100 metric 100 2026-03-10T10:14:42.500 INFO:teuthology.orchestra.run.vm00.stdout:/usr/sbin/ip: stdout 192.168.123.1 dev ens3 proto dhcp scope link src 192.168.123.100 metric 100 2026-03-10T10:14:42.501 INFO:teuthology.orchestra.run.vm00.stdout:/usr/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-10T10:14:42.501 INFO:teuthology.orchestra.run.vm00.stdout:/usr/sbin/ip: stdout fe80::/64 dev ens3 proto kernel metric 256 pref medium 2026-03-10T10:14:42.503 INFO:teuthology.orchestra.run.vm00.stdout:/usr/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-10T10:14:42.503 INFO:teuthology.orchestra.run.vm00.stdout:/usr/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-10T10:14:42.503 INFO:teuthology.orchestra.run.vm00.stdout:/usr/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-10T10:14:42.503 INFO:teuthology.orchestra.run.vm00.stdout:/usr/sbin/ip: stdout 2: ens3: mtu 1500 state UP qlen 1000 2026-03-10T10:14:42.503 INFO:teuthology.orchestra.run.vm00.stdout:/usr/sbin/ip: stdout inet6 fe80::5055:ff:fe00:0/64 scope link 2026-03-10T10:14:42.503 INFO:teuthology.orchestra.run.vm00.stdout:/usr/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-10T10:14:42.503 INFO:teuthology.orchestra.run.vm00.stdout:Mon IP `192.168.123.100` is in CIDR network `192.168.123.0/24` 2026-03-10T10:14:42.503 INFO:teuthology.orchestra.run.vm00.stdout:Mon IP `192.168.123.100` is in CIDR network `192.168.123.0/24` 2026-03-10T10:14:42.503 INFO:teuthology.orchestra.run.vm00.stdout:Mon IP `192.168.123.100` is in CIDR network `192.168.123.1/32` 2026-03-10T10:14:42.503 INFO:teuthology.orchestra.run.vm00.stdout:Mon IP `192.168.123.100` is in CIDR network `192.168.123.1/32` 2026-03-10T10:14:42.503 INFO:teuthology.orchestra.run.vm00.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24', '192.168.123.1/32', '192.168.123.1/32'] 2026-03-10T10:14:42.503 INFO:teuthology.orchestra.run.vm00.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-10T10:14:42.504 INFO:teuthology.orchestra.run.vm00.stdout:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T10:14:43.551 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/docker: stdout e911bdebe5c8faa3800735d1568fcdca65db60df: Pulling from ceph-ci/ceph 2026-03-10T10:14:43.551 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/docker: stdout Digest: sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T10:14:43.551 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/docker: stdout Status: Image is up to date for quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T10:14:43.551 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/docker: stdout quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T10:14:43.716 INFO:teuthology.orchestra.run.vm00.stdout:ceph: stdout ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-10T10:14:43.716 INFO:teuthology.orchestra.run.vm00.stdout:Ceph version: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-10T10:14:43.716 INFO:teuthology.orchestra.run.vm00.stdout:Extracting ceph user uid/gid from container image... 2026-03-10T10:14:43.861 INFO:teuthology.orchestra.run.vm00.stdout:stat: stdout 167 167 2026-03-10T10:14:43.861 INFO:teuthology.orchestra.run.vm00.stdout:Creating initial keys... 2026-03-10T10:14:43.953 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-authtool: stdout AQAT769pMSVTNxAAvewOmBrZFk++iizRKS4UWQ== 2026-03-10T10:14:44.053 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-authtool: stdout AQAU769pW8KxARAAcFGqaTH56V2VGvWvga6YYQ== 2026-03-10T10:14:44.165 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-authtool: stdout AQAU769pAFcvBxAAR736fwf7NkzC7fF6ZiQumQ== 2026-03-10T10:14:44.165 INFO:teuthology.orchestra.run.vm00.stdout:Creating initial monmap... 2026-03-10T10:14:44.283 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-10T10:14:44.283 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-10T10:14:44.283 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:14:44.283 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-10T10:14:44.283 INFO:teuthology.orchestra.run.vm00.stdout:monmaptool for a [v2:192.168.123.100:3300,v1:192.168.123.100:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-10T10:14:44.283 INFO:teuthology.orchestra.run.vm00.stdout:setting min_mon_release = quincy 2026-03-10T10:14:44.283 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: set fsid to 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:14:44.283 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-10T10:14:44.283 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:14:44.283 INFO:teuthology.orchestra.run.vm00.stdout:Creating mon... 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 1 imported monmap: 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr epoch 0 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr last_changed 2026-03-10T10:14:44.248116+0000 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr min_mon_release 17 (quincy) 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr election_strategy: 1 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 0 /usr/bin/ceph-mon: set fsid to 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Git sha 0 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: DB SUMMARY 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: DB Session ID: N2U7L37M7QZJWU1WJEPI 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 0, files: 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.error_if_exists: 0 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.create_if_missing: 1 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.env: 0x55d00f064dc0 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.info_log: 0x55d0269b4da0 2026-03-10T10:14:44.424 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.statistics: (nil) 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.use_fsync: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.db_log_dir: 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.wal_dir: 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.write_buffer_manager: 0x55d0269ab5e0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.unordered_write: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.row_cache: None 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.wal_filter: None 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.two_write_queues: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.wal_compression: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.atomic_flush: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-10T10:14:44.425 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.max_open_files: -1 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Compression algorithms supported: 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: kZSTD supported: 0 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: kXpressCompression supported: 0 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: kZlibCompression supported: 1 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.373+0000 7f893c650d80 4 rocksdb: [db/db_impl/db_impl_open.cc:317] Creating manifest 1 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000001 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.merge_operator: 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compaction_filter: None 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55d0269a7520) 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr cache_index_and_filter_blocks: 1 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr cache_index_and_filter_blocks_with_high_priority: 0 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr pin_top_level_index_and_filter: 1 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr index_type: 0 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr data_block_index_type: 0 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr index_shortening: 1 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr data_block_hash_table_util_ratio: 0.750000 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr checksum: 4 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr no_block_cache: 0 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr block_cache: 0x55d0269cd350 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr block_cache_name: BinnedLRUCache 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr block_cache_options: 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr capacity : 536870912 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr num_shard_bits : 4 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr strict_capacity_limit : 0 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr high_pri_pool_ratio: 0.000 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr block_cache_compressed: (nil) 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr persistent_cache: (nil) 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr block_size: 4096 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr block_size_deviation: 10 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr block_restart_interval: 16 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr index_block_restart_interval: 1 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr metadata_block_size: 4096 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr partition_filters: 0 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr use_delta_encoding: 1 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr filter_policy: bloomfilter 2026-03-10T10:14:44.426 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr whole_key_filtering: 1 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr verify_compression: 0 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr read_amp_bytes_per_bit: 0 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr format_version: 5 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr enable_index_compression: 1 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr block_align: 0 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr max_auto_readahead_size: 262144 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr prepopulate_block_cache: 0 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr initial_auto_readahead_size: 8192 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr num_file_reads_for_auto_readahead: 2 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compression: NoCompression 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.num_levels: 7 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-10T10:14:44.427 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.bloom_locality: 0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.ttl: 2592000 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.enable_blob_files: false 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.min_blob_size: 0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000001 succeeded,manifest_file_number is 1, next_file_number is 3, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 47b673dc-67c7-48f5-9b24-9515b0cbe20b 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.377+0000 7f893c650d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 5 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.381+0000 7f893c650d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55d0269cee00 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.381+0000 7f893c650d80 4 rocksdb: DB pointer 0x55d026ab2000 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.381+0000 7f8933dda640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.381+0000 7f8933dda640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr ** DB Stats ** 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr ** Compaction Stats [default] ** 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr ** Compaction Stats [default] ** 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T10:14:44.428 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Flush(GB): cumulative 0.000, interval 0.000 2026-03-10T10:14:44.429 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr AddFile(GB): cumulative 0.000, interval 0.000 2026-03-10T10:14:44.429 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr AddFile(Total Files): cumulative 0, interval 0 2026-03-10T10:14:44.429 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr AddFile(L0 Files): cumulative 0, interval 0 2026-03-10T10:14:44.429 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr AddFile(Keys): cumulative 0, interval 0 2026-03-10T10:14:44.429 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T10:14:44.429 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T10:14:44.429 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-10T10:14:44.429 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Block cache BinnedLRUCache@0x55d0269cd350#7 capacity: 512.00 MB usage: 0.00 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 7e-06 secs_since: 0 2026-03-10T10:14:44.429 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr Block cache entry stats(count,size,portion): Misc(1,0.00 KB,0%) 2026-03-10T10:14:44.429 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.429 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr ** File Read Latency Histogram By Level [default] ** 2026-03-10T10:14:44.429 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr 2026-03-10T10:14:44.429 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.381+0000 7f893c650d80 4 rocksdb: [db/db_impl/db_impl.cc:496] Shutdown: canceling all background work 2026-03-10T10:14:44.429 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.381+0000 7f893c650d80 4 rocksdb: [db/db_impl/db_impl.cc:704] Shutdown complete 2026-03-10T10:14:44.429 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-10T10:14:44.381+0000 7f893c650d80 0 /usr/bin/ceph-mon: created monfs at /var/lib/ceph/mon/ceph-a for mon.a 2026-03-10T10:14:44.429 INFO:teuthology.orchestra.run.vm00.stdout:create mon.a on 2026-03-10T10:14:44.589 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Removed /etc/systemd/system/multi-user.target.wants/ceph.target. 2026-03-10T10:14:44.743 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-10T10:14:44.912 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad.target → /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad.target. 2026-03-10T10:14:44.912 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad.target → /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad.target. 2026-03-10T10:14:45.086 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mon.a 2026-03-10T10:14:45.086 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Failed to reset failed state of unit ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mon.a.service: Unit ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mon.a.service not loaded. 2026-03-10T10:14:45.251 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad.target.wants/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mon.a.service → /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service. 2026-03-10T10:14:45.258 INFO:teuthology.orchestra.run.vm00.stdout:firewalld does not appear to be present 2026-03-10T10:14:45.259 INFO:teuthology.orchestra.run.vm00.stdout:Not possible to enable service . firewalld.service is not available 2026-03-10T10:14:45.259 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mon to start... 2026-03-10T10:14:45.259 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mon... 2026-03-10T10:14:45.467 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:45 vm00 bash[20239]: cluster 2026-03-10T10:14:45.384657+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T10:14:45.492 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout cluster: 2026-03-10T10:14:45.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout id: 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:14:45.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-10T10:14:45.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-10T10:14:45.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout services: 2026-03-10T10:14:45.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum a (age 0.0701139s) 2026-03-10T10:14:45.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-10T10:14:45.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-10T10:14:45.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-10T10:14:45.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout data: 2026-03-10T10:14:45.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-10T10:14:45.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-10T10:14:45.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-10T10:14:45.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout pgs: 2026-03-10T10:14:45.493 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-10T10:14:45.493 INFO:teuthology.orchestra.run.vm00.stdout:mon is available 2026-03-10T10:14:45.493 INFO:teuthology.orchestra.run.vm00.stdout:Assimilating anything we can from ceph.conf... 2026-03-10T10:14:45.678 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-10T10:14:45.679 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [global] 2026-03-10T10:14:45.679 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout fsid = 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:14:45.679 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-10T10:14:45.679 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.100:3300,v1:192.168.123.100:6789] 2026-03-10T10:14:45.679 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-10T10:14:45.679 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-10T10:14:45.679 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-10T10:14:45.679 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-10T10:14:45.679 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-10T10:14:45.679 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-10T10:14:45.679 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mgr/cephadm/use_agent = True 2026-03-10T10:14:45.679 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-10T10:14:45.679 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-10T10:14:45.679 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [osd] 2026-03-10T10:14:45.679 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-10T10:14:45.679 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-10T10:14:45.679 INFO:teuthology.orchestra.run.vm00.stdout:Generating new minimal ceph.conf... 2026-03-10T10:14:45.853 INFO:teuthology.orchestra.run.vm00.stdout:Restarting the monitor... 2026-03-10T10:14:46.034 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:45 vm00 systemd[1]: Stopping Ceph mon.a for 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad... 2026-03-10T10:14:46.034 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:45 vm00 bash[20239]: debug 2026-03-10T10:14:45.885+0000 7ff0ea2e9640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-10T10:14:46.034 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:45 vm00 bash[20239]: debug 2026-03-10T10:14:45.885+0000 7ff0ea2e9640 -1 mon.a@0(leader) e1 *** Got Signal Terminated *** 2026-03-10T10:14:46.084 INFO:teuthology.orchestra.run.vm00.stdout:Setting public_network to 192.168.123.1/32,192.168.123.0/24 in mon config section 2026-03-10T10:14:46.295 INFO:teuthology.orchestra.run.vm00.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-10T10:14:46.295 INFO:teuthology.orchestra.run.vm00.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:14:46.295 INFO:teuthology.orchestra.run.vm00.stdout:Creating mgr... 2026-03-10T10:14:46.295 INFO:teuthology.orchestra.run.vm00.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-10T10:14:46.296 INFO:teuthology.orchestra.run.vm00.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20624]: ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad-mon-a 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 systemd[1]: ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mon.a.service: Deactivated successfully. 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 systemd[1]: Stopped Ceph mon.a for 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad. 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 systemd[1]: Started Ceph mon.a for 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad. 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.197+0000 7ff3ce477d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.197+0000 7ff3ce477d80 0 ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 7 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.197+0000 7ff3ce477d80 0 pidfile_write: ignore empty --pid-file 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 0 load: jerasure load: lrc 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Git sha 0 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: DB SUMMARY 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: DB Session ID: GH6939SI22SZKRXCT0BU 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: CURRENT file: CURRENT 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 1, files: 000008.sst 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000009.log size: 75507 ; 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.error_if_exists: 0 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.create_if_missing: 0 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.env: 0x558ab51ecdc0 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.info_log: 0x558ad25ac700 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.statistics: (nil) 2026-03-10T10:14:46.306 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.use_fsync: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.db_log_dir: 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.wal_dir: 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.write_buffer_manager: 0x558ad25b1900 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.unordered_write: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.row_cache: None 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.wal_filter: None 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.two_write_queues: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.wal_compression: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.atomic_flush: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-10T10:14:46.307 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_open_files: -1 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Compression algorithms supported: 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: kZSTD supported: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: kXpressCompression supported: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: kZlibCompression supported: 1 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.merge_operator: 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compaction_filter: None 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x558ad25ac640) 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cache_index_and_filter_blocks: 1 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: pin_top_level_index_and_filter: 1 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: index_type: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: data_block_index_type: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: index_shortening: 1 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: data_block_hash_table_util_ratio: 0.750000 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: checksum: 4 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: no_block_cache: 0 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: block_cache: 0x558ad25d3350 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: block_cache_name: BinnedLRUCache 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: block_cache_options: 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: capacity : 536870912 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: num_shard_bits : 4 2026-03-10T10:14:46.308 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: strict_capacity_limit : 0 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: high_pri_pool_ratio: 0.000 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: block_cache_compressed: (nil) 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: persistent_cache: (nil) 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: block_size: 4096 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: block_size_deviation: 10 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: block_restart_interval: 16 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: index_block_restart_interval: 1 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: metadata_block_size: 4096 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: partition_filters: 0 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: use_delta_encoding: 1 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: filter_policy: bloomfilter 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: whole_key_filtering: 1 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: verify_compression: 0 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: read_amp_bytes_per_bit: 0 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: format_version: 5 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: enable_index_compression: 1 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: block_align: 0 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: max_auto_readahead_size: 262144 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: prepopulate_block_cache: 0 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: initial_auto_readahead_size: 8192 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: num_file_reads_for_auto_readahead: 2 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compression: NoCompression 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.num_levels: 7 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-10T10:14:46.309 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.bloom_locality: 0 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.ttl: 2592000 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.enable_blob_files: false 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.min_blob_size: 0 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 47b673dc-67c7-48f5-9b24-9515b0cbe20b 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773137686204272, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773137686205796, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 72588, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 225, "table_properties": {"data_size": 70867, "index_size": 174, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 517, "raw_key_size": 9705, "raw_average_key_size": 49, "raw_value_size": 65346, "raw_average_value_size": 333, "num_data_blocks": 8, "num_entries": 196, "num_filter_entries": 196, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773137686, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "47b673dc-67c7-48f5-9b24-9515b0cbe20b", "db_session_id": "GH6939SI22SZKRXCT0BU", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773137686205846, "job": 1, "event": "recovery_finished"} 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: debug 2026-03-10T10:14:46.201+0000 7ff3ce477d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.219857+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.219857+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.219882+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.219882+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-10T10:14:46.310 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.219886+0000 mon.a (mon.0) 3 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.219886+0000 mon.a (mon.0) 3 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.219889+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-10T10:14:44.248116+0000 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.219889+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-10T10:14:44.248116+0000 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.219895+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.219895+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.219898+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.219898+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.219902+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.219902+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.219905+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.219905+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.220108+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.220108+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.220117+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.220117+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.220561+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: cluster 2026-03-10T10:14:46.220561+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: audit 2026-03-10T10:14:46.258572+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.100:0/624538143' entity='client.admin' 2026-03-10T10:14:46.311 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20709]: audit 2026-03-10T10:14:46.258572+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.100:0/624538143' entity='client.admin' 2026-03-10T10:14:46.456 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mgr.a 2026-03-10T10:14:46.456 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Failed to reset failed state of unit ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mgr.a.service: Unit ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mgr.a.service not loaded. 2026-03-10T10:14:46.558 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:14:46.603 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad.target.wants/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mgr.a.service → /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service. 2026-03-10T10:14:46.610 INFO:teuthology.orchestra.run.vm00.stdout:firewalld does not appear to be present 2026-03-10T10:14:46.611 INFO:teuthology.orchestra.run.vm00.stdout:Not possible to enable service . firewalld.service is not available 2026-03-10T10:14:46.611 INFO:teuthology.orchestra.run.vm00.stdout:firewalld does not appear to be present 2026-03-10T10:14:46.611 INFO:teuthology.orchestra.run.vm00.stdout:Not possible to open ports <[9283, 8765]>. firewalld.service is not available 2026-03-10T10:14:46.611 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mgr to start... 2026-03-10T10:14:46.611 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mgr... 2026-03-10T10:14:46.812 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:46 vm00 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:14:46.826 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-10T10:14:46.826 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-10T10:14:46.826 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsid": "9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad", 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 0 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "a" 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_age": 0, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T10:14:45:388655+0000", 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T10:14:46.827 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T10:14:45.389320+0000", 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-10T10:14:46.828 INFO:teuthology.orchestra.run.vm00.stdout:mgr not available, waiting (1/15)... 2026-03-10T10:14:47.168 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:46 vm00 bash[20974]: debug 2026-03-10T10:14:46.949+0000 7f51e7b1d140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T10:14:47.603 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:47 vm00 bash[20709]: audit 2026-03-10T10:14:46.778814+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.100:0/2079901906' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:14:47.604 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:47 vm00 bash[20709]: audit 2026-03-10T10:14:46.778814+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.100:0/2079901906' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:14:47.604 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:47 vm00 bash[20974]: debug 2026-03-10T10:14:47.233+0000 7f51e7b1d140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T10:14:47.856 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:47 vm00 bash[20974]: debug 2026-03-10T10:14:47.657+0000 7f51e7b1d140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T10:14:47.856 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:47 vm00 bash[20974]: debug 2026-03-10T10:14:47.737+0000 7f51e7b1d140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T10:14:47.857 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:47 vm00 bash[20974]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T10:14:47.857 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:47 vm00 bash[20974]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T10:14:47.857 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:47 vm00 bash[20974]: from numpy import show_config as show_numpy_config 2026-03-10T10:14:47.857 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:47 vm00 bash[20974]: debug 2026-03-10T10:14:47.853+0000 7f51e7b1d140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T10:14:48.144 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:47 vm00 bash[20974]: debug 2026-03-10T10:14:47.977+0000 7f51e7b1d140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T10:14:48.144 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:48 vm00 bash[20974]: debug 2026-03-10T10:14:48.013+0000 7f51e7b1d140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T10:14:48.144 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:48 vm00 bash[20974]: debug 2026-03-10T10:14:48.049+0000 7f51e7b1d140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T10:14:48.144 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:48 vm00 bash[20974]: debug 2026-03-10T10:14:48.093+0000 7f51e7b1d140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T10:14:48.144 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:48 vm00 bash[20974]: debug 2026-03-10T10:14:48.141+0000 7f51e7b1d140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T10:14:48.814 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:48 vm00 bash[20974]: debug 2026-03-10T10:14:48.553+0000 7f51e7b1d140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T10:14:48.814 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:48 vm00 bash[20974]: debug 2026-03-10T10:14:48.589+0000 7f51e7b1d140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T10:14:48.814 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:48 vm00 bash[20974]: debug 2026-03-10T10:14:48.625+0000 7f51e7b1d140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T10:14:48.814 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:48 vm00 bash[20974]: debug 2026-03-10T10:14:48.765+0000 7f51e7b1d140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T10:14:48.814 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:48 vm00 bash[20974]: debug 2026-03-10T10:14:48.809+0000 7f51e7b1d140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsid": "9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad", 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 0 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "a" 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_age": 2, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T10:14:49.061 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T10:14:45:388655+0000", 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T10:14:45.389320+0000", 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-10T10:14:49.062 INFO:teuthology.orchestra.run.vm00.stdout:mgr not available, waiting (2/15)... 2026-03-10T10:14:49.158 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:49 vm00 bash[20709]: audit 2026-03-10T10:14:49.008480+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.100:0/3321573109' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:14:49.158 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:49 vm00 bash[20709]: audit 2026-03-10T10:14:49.008480+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.100:0/3321573109' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:14:49.158 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:48 vm00 bash[20974]: debug 2026-03-10T10:14:48.857+0000 7f51e7b1d140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T10:14:49.158 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:48 vm00 bash[20974]: debug 2026-03-10T10:14:48.985+0000 7f51e7b1d140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T10:14:49.415 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:49 vm00 bash[20974]: debug 2026-03-10T10:14:49.153+0000 7f51e7b1d140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T10:14:49.415 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:49 vm00 bash[20974]: debug 2026-03-10T10:14:49.333+0000 7f51e7b1d140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T10:14:49.415 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:49 vm00 bash[20974]: debug 2026-03-10T10:14:49.365+0000 7f51e7b1d140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T10:14:49.668 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:49 vm00 bash[20974]: debug 2026-03-10T10:14:49.409+0000 7f51e7b1d140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T10:14:49.668 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:49 vm00 bash[20974]: debug 2026-03-10T10:14:49.609+0000 7f51e7b1d140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T10:14:50.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: cluster 2026-03-10T10:14:49.964922+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon a 2026-03-10T10:14:50.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: cluster 2026-03-10T10:14:49.964922+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon a 2026-03-10T10:14:50.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: cluster 2026-03-10T10:14:49.971198+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: a(active, starting, since 0.00642738s) 2026-03-10T10:14:50.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: cluster 2026-03-10T10:14:49.971198+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: a(active, starting, since 0.00642738s) 2026-03-10T10:14:50.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:49.973767+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:14:50.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:49.973767+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:14:50.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:49.974443+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:14:50.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:49.974443+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:14:50.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:49.974506+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:14:50.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:49.974506+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:14:50.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:49.974553+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:14:50.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:49.974553+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:14:50.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:49.974598+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:14:50.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:49.974598+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:14:50.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: cluster 2026-03-10T10:14:49.983830+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon a is now available 2026-03-10T10:14:50.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: cluster 2026-03-10T10:14:49.983830+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon a is now available 2026-03-10T10:14:50.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:49.997307+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:14:50.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:49.997307+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:14:50.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:49.999174+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:14:50.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:49.999174+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:14:50.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:50.002748+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:14:50.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:50.002748+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:14:50.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:50.005327+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:14:50.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:50.005327+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:14:50.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:50.006065+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:14:50.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:50 vm00 bash[20709]: audit 2026-03-10T10:14:50.006065+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:14:50.169 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:49 vm00 bash[20974]: debug 2026-03-10T10:14:49.961+0000 7f51e7b1d140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsid": "9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad", 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 0 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "a" 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T10:14:51.418 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T10:14:45:388655+0000", 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T10:14:45.389320+0000", 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-10T10:14:51.419 INFO:teuthology.orchestra.run.vm00.stdout:mgr is available 2026-03-10T10:14:51.663 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-10T10:14:51.663 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [global] 2026-03-10T10:14:51.664 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout fsid = 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:14:51.664 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-10T10:14:51.664 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.100:3300,v1:192.168.123.100:6789] 2026-03-10T10:14:51.664 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-10T10:14:51.664 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-10T10:14:51.664 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-10T10:14:51.664 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-10T10:14:51.664 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-10T10:14:51.664 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-10T10:14:51.664 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-10T10:14:51.664 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-10T10:14:51.664 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [osd] 2026-03-10T10:14:51.664 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-10T10:14:51.664 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-10T10:14:51.664 INFO:teuthology.orchestra.run.vm00.stdout:Enabling cephadm module... 2026-03-10T10:14:52.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:51 vm00 bash[20709]: cluster 2026-03-10T10:14:50.977179+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: a(active, since 1.0124s) 2026-03-10T10:14:52.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:51 vm00 bash[20709]: cluster 2026-03-10T10:14:50.977179+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: a(active, since 1.0124s) 2026-03-10T10:14:52.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:51 vm00 bash[20709]: audit 2026-03-10T10:14:51.382436+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.100:0/540976360' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:14:52.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:51 vm00 bash[20709]: audit 2026-03-10T10:14:51.382436+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.100:0/540976360' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:14:52.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:51 vm00 bash[20709]: audit 2026-03-10T10:14:51.619140+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.100:0/3131372554' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T10:14:52.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:51 vm00 bash[20709]: audit 2026-03-10T10:14:51.619140+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.100:0/3131372554' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T10:14:52.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:51 vm00 bash[20709]: audit 2026-03-10T10:14:51.622788+0000 mon.a (mon.0) 31 : audit [INF] from='client.? 192.168.123.100:0/3131372554' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-10T10:14:52.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:51 vm00 bash[20709]: audit 2026-03-10T10:14:51.622788+0000 mon.a (mon.0) 31 : audit [INF] from='client.? 192.168.123.100:0/3131372554' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-10T10:14:52.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:51 vm00 bash[20709]: audit 2026-03-10T10:14:51.880558+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.100:0/1053282729' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T10:14:52.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:51 vm00 bash[20709]: audit 2026-03-10T10:14:51.880558+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.100:0/1053282729' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T10:14:52.918 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:52 vm00 bash[20974]: ignoring --setuser ceph since I am not root 2026-03-10T10:14:52.918 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:52 vm00 bash[20974]: ignoring --setgroup ceph since I am not root 2026-03-10T10:14:52.918 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:52 vm00 bash[20974]: debug 2026-03-10T10:14:52.773+0000 7fd3a156e140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T10:14:52.918 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:52 vm00 bash[20974]: debug 2026-03-10T10:14:52.809+0000 7fd3a156e140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T10:14:53.002 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-10T10:14:53.002 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 4, 2026-03-10T10:14:53.002 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T10:14:53.002 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-10T10:14:53.002 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-10T10:14:53.002 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-10T10:14:53.002 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for the mgr to restart... 2026-03-10T10:14:53.002 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mgr epoch 4... 2026-03-10T10:14:53.254 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:52 vm00 bash[20974]: debug 2026-03-10T10:14:52.941+0000 7fd3a156e140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T10:14:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:53 vm00 bash[20709]: audit 2026-03-10T10:14:52.624524+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.100:0/1053282729' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T10:14:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:53 vm00 bash[20709]: audit 2026-03-10T10:14:52.624524+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.100:0/1053282729' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T10:14:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:53 vm00 bash[20709]: cluster 2026-03-10T10:14:52.626178+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-10T10:14:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:53 vm00 bash[20709]: cluster 2026-03-10T10:14:52.626178+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-10T10:14:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:53 vm00 bash[20709]: audit 2026-03-10T10:14:52.937791+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.100:0/3841211786' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T10:14:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:53 vm00 bash[20709]: audit 2026-03-10T10:14:52.937791+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.100:0/3841211786' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T10:14:53.918 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:53 vm00 bash[20974]: debug 2026-03-10T10:14:53.249+0000 7fd3a156e140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T10:14:53.918 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:53 vm00 bash[20974]: debug 2026-03-10T10:14:53.665+0000 7fd3a156e140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T10:14:53.918 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:53 vm00 bash[20974]: debug 2026-03-10T10:14:53.745+0000 7fd3a156e140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T10:14:53.918 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:53 vm00 bash[20974]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T10:14:53.918 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:53 vm00 bash[20974]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T10:14:53.918 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:53 vm00 bash[20974]: from numpy import show_config as show_numpy_config 2026-03-10T10:14:53.918 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:53 vm00 bash[20974]: debug 2026-03-10T10:14:53.865+0000 7fd3a156e140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T10:14:54.418 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:54 vm00 bash[20974]: debug 2026-03-10T10:14:54.001+0000 7fd3a156e140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T10:14:54.418 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:54 vm00 bash[20974]: debug 2026-03-10T10:14:54.037+0000 7fd3a156e140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T10:14:54.418 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:54 vm00 bash[20974]: debug 2026-03-10T10:14:54.069+0000 7fd3a156e140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T10:14:54.418 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:54 vm00 bash[20974]: debug 2026-03-10T10:14:54.109+0000 7fd3a156e140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T10:14:54.418 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:54 vm00 bash[20974]: debug 2026-03-10T10:14:54.157+0000 7fd3a156e140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T10:14:54.832 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:54 vm00 bash[20974]: debug 2026-03-10T10:14:54.553+0000 7fd3a156e140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T10:14:54.833 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:54 vm00 bash[20974]: debug 2026-03-10T10:14:54.585+0000 7fd3a156e140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T10:14:54.833 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:54 vm00 bash[20974]: debug 2026-03-10T10:14:54.621+0000 7fd3a156e140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T10:14:54.833 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:54 vm00 bash[20974]: debug 2026-03-10T10:14:54.753+0000 7fd3a156e140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T10:14:54.833 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:54 vm00 bash[20974]: debug 2026-03-10T10:14:54.793+0000 7fd3a156e140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T10:14:55.084 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:54 vm00 bash[20974]: debug 2026-03-10T10:14:54.829+0000 7fd3a156e140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T10:14:55.085 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:54 vm00 bash[20974]: debug 2026-03-10T10:14:54.933+0000 7fd3a156e140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T10:14:55.418 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20974]: debug 2026-03-10T10:14:55.081+0000 7fd3a156e140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T10:14:55.418 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20974]: debug 2026-03-10T10:14:55.241+0000 7fd3a156e140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T10:14:55.418 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20974]: debug 2026-03-10T10:14:55.273+0000 7fd3a156e140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T10:14:55.418 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20974]: debug 2026-03-10T10:14:55.313+0000 7fd3a156e140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T10:14:55.721 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20974]: debug 2026-03-10T10:14:55.453+0000 7fd3a156e140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T10:14:55.721 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20974]: debug 2026-03-10T10:14:55.661+0000 7fd3a156e140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T10:14:56.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: cluster 2026-03-10T10:14:55.665147+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon a restarted 2026-03-10T10:14:56.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: cluster 2026-03-10T10:14:55.665147+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon a restarted 2026-03-10T10:14:56.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: cluster 2026-03-10T10:14:55.665541+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon a 2026-03-10T10:14:56.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: cluster 2026-03-10T10:14:55.665541+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon a 2026-03-10T10:14:56.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: cluster 2026-03-10T10:14:55.669966+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-10T10:14:56.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: cluster 2026-03-10T10:14:55.669966+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-10T10:14:56.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: cluster 2026-03-10T10:14:55.670031+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e5: a(active, starting, since 0.00457834s) 2026-03-10T10:14:56.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: cluster 2026-03-10T10:14:55.670031+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e5: a(active, starting, since 0.00457834s) 2026-03-10T10:14:56.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.674294+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:14:56.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.674294+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:14:56.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.674913+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:14:56.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.674913+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:14:56.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.675645+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:14:56.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.675645+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:14:56.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.675986+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:14:56.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.675986+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:14:56.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.676303+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:14:56.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.676303+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:14:56.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: cluster 2026-03-10T10:14:55.681445+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon a is now available 2026-03-10T10:14:56.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: cluster 2026-03-10T10:14:55.681445+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon a is now available 2026-03-10T10:14:56.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.689780+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:14:56.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.689780+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:14:56.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.692910+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:14:56.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.692910+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:14:56.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.695465+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:14:56.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.695465+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:14:56.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.711238+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:14:56.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.711238+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:14:56.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.713169+0000 mon.a (mon.0) 50 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:14:56.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:55 vm00 bash[20709]: audit 2026-03-10T10:14:55.713169+0000 mon.a (mon.0) 50 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:14:56.717 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-10T10:14:56.718 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 6, 2026-03-10T10:14:56.718 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-10T10:14:56.718 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-10T10:14:56.718 INFO:teuthology.orchestra.run.vm00.stdout:mgr epoch 4 is available 2026-03-10T10:14:56.718 INFO:teuthology.orchestra.run.vm00.stdout:Setting orchestrator backend to cephadm... 2026-03-10T10:14:57.008 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:56 vm00 bash[20709]: cephadm 2026-03-10T10:14:55.687288+0000 mgr.a (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-10T10:14:57.008 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:56 vm00 bash[20709]: cephadm 2026-03-10T10:14:55.687288+0000 mgr.a (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-10T10:14:57.008 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:56 vm00 bash[20709]: audit 2026-03-10T10:14:55.717231+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:14:57.008 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:56 vm00 bash[20709]: audit 2026-03-10T10:14:55.717231+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:14:57.008 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:56 vm00 bash[20709]: audit 2026-03-10T10:14:55.730672+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:14:57.008 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:56 vm00 bash[20709]: audit 2026-03-10T10:14:55.730672+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:14:57.008 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:56 vm00 bash[20709]: audit 2026-03-10T10:14:56.084022+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:14:57.009 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:56 vm00 bash[20709]: audit 2026-03-10T10:14:56.084022+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:14:57.009 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:56 vm00 bash[20709]: audit 2026-03-10T10:14:56.086584+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:14:57.009 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:56 vm00 bash[20709]: audit 2026-03-10T10:14:56.086584+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:14:57.009 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:56 vm00 bash[20709]: cephadm 2026-03-10T10:14:56.600038+0000 mgr.a (mgr.14118) 2 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Bus STARTING 2026-03-10T10:14:57.009 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:56 vm00 bash[20709]: cephadm 2026-03-10T10:14:56.600038+0000 mgr.a (mgr.14118) 2 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Bus STARTING 2026-03-10T10:14:57.009 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:56 vm00 bash[20709]: cluster 2026-03-10T10:14:56.673593+0000 mon.a (mon.0) 55 : cluster [DBG] mgrmap e6: a(active, since 1.00814s) 2026-03-10T10:14:57.009 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:56 vm00 bash[20709]: cluster 2026-03-10T10:14:56.673593+0000 mon.a (mon.0) 55 : cluster [DBG] mgrmap e6: a(active, since 1.00814s) 2026-03-10T10:14:57.009 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:56 vm00 bash[20709]: audit 2026-03-10T10:14:56.675529+0000 mgr.a (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T10:14:57.009 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:56 vm00 bash[20709]: audit 2026-03-10T10:14:56.675529+0000 mgr.a (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T10:14:57.287 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-10T10:14:57.287 INFO:teuthology.orchestra.run.vm00.stdout:Generating ssh key... 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:56.679528+0000 mgr.a (mgr.14118) 4 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:56.679528+0000 mgr.a (mgr.14118) 4 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: cephadm 2026-03-10T10:14:56.707900+0000 mgr.a (mgr.14118) 5 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Serving on https://192.168.123.100:7150 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: cephadm 2026-03-10T10:14:56.707900+0000 mgr.a (mgr.14118) 5 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Serving on https://192.168.123.100:7150 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: cephadm 2026-03-10T10:14:56.708300+0000 mgr.a (mgr.14118) 6 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Client ('192.168.123.100', 49166) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: cephadm 2026-03-10T10:14:56.708300+0000 mgr.a (mgr.14118) 6 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Client ('192.168.123.100', 49166) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: cephadm 2026-03-10T10:14:56.808912+0000 mgr.a (mgr.14118) 7 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Serving on http://192.168.123.100:8765 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: cephadm 2026-03-10T10:14:56.808912+0000 mgr.a (mgr.14118) 7 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Serving on http://192.168.123.100:8765 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: cephadm 2026-03-10T10:14:56.809134+0000 mgr.a (mgr.14118) 8 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Bus STARTED 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: cephadm 2026-03-10T10:14:56.809134+0000 mgr.a (mgr.14118) 8 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Bus STARTED 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:56.809685+0000 mon.a (mon.0) 56 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:56.809685+0000 mon.a (mon.0) 56 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:56.988109+0000 mgr.a (mgr.14118) 9 : audit [DBG] from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:56.988109+0000 mgr.a (mgr.14118) 9 : audit [DBG] from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:56.992162+0000 mon.a (mon.0) 57 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:56.992162+0000 mon.a (mon.0) 57 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:56.999054+0000 mon.a (mon.0) 58 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:56.999054+0000 mon.a (mon.0) 58 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:57.253328+0000 mgr.a (mgr.14118) 10 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:57.253328+0000 mgr.a (mgr.14118) 10 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:57.525895+0000 mgr.a (mgr.14118) 11 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:57.525895+0000 mgr.a (mgr.14118) 11 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: cephadm 2026-03-10T10:14:57.526082+0000 mgr.a (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: cephadm 2026-03-10T10:14:57.526082+0000 mgr.a (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:57.541819+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:57.541819+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:57.543843+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:14:57.790 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20709]: audit 2026-03-10T10:14:57.543843+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:14:57.791 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20974]: Generating public/private ed25519 key pair. 2026-03-10T10:14:57.791 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20974]: Your identification has been saved in /tmp/tmp6gxmz5bt/key 2026-03-10T10:14:57.791 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20974]: Your public key has been saved in /tmp/tmp6gxmz5bt/key.pub 2026-03-10T10:14:57.791 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20974]: The key fingerprint is: 2026-03-10T10:14:57.791 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20974]: SHA256:N5w+XSSBayRfecZF2NaWADxY/TkQ2EP+wOWANcMFO5M ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:14:57.791 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20974]: The key's randomart image is: 2026-03-10T10:14:57.791 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20974]: +--[ED25519 256]--+ 2026-03-10T10:14:57.791 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20974]: | ++@%=B=| 2026-03-10T10:14:57.791 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20974]: | ..o==B&.=| 2026-03-10T10:14:57.791 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20974]: | + ooE=+.| 2026-03-10T10:14:57.791 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20974]: | .+. o=+ | 2026-03-10T10:14:57.791 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20974]: | S.= ...| 2026-03-10T10:14:57.791 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20974]: | o o . | 2026-03-10T10:14:57.791 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20974]: | o . | 2026-03-10T10:14:57.791 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20974]: | . | 2026-03-10T10:14:57.791 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20974]: | | 2026-03-10T10:14:57.791 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:14:57 vm00 bash[20974]: +----[SHA256]-----+ 2026-03-10T10:14:57.813 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIKXzt4Lk4e3wujrSBUxsCaFaFvtK1ABc232NZhzsYY1G ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:14:57.813 INFO:teuthology.orchestra.run.vm00.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-10T10:14:57.813 INFO:teuthology.orchestra.run.vm00.stdout:Adding key to root@localhost authorized_keys... 2026-03-10T10:14:57.813 INFO:teuthology.orchestra.run.vm00.stdout:Adding host vm00... 2026-03-10T10:14:59.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:59 vm00 bash[20709]: audit 2026-03-10T10:14:57.778687+0000 mgr.a (mgr.14118) 13 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:14:59.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:59 vm00 bash[20709]: audit 2026-03-10T10:14:57.778687+0000 mgr.a (mgr.14118) 13 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:14:59.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:59 vm00 bash[20709]: audit 2026-03-10T10:14:58.023626+0000 mgr.a (mgr.14118) 14 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm00", "addr": "192.168.123.100", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:14:59.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:59 vm00 bash[20709]: audit 2026-03-10T10:14:58.023626+0000 mgr.a (mgr.14118) 14 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm00", "addr": "192.168.123.100", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:14:59.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:59 vm00 bash[20709]: cluster 2026-03-10T10:14:58.548356+0000 mon.a (mon.0) 61 : cluster [DBG] mgrmap e7: a(active, since 2s) 2026-03-10T10:14:59.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:14:59 vm00 bash[20709]: cluster 2026-03-10T10:14:58.548356+0000 mon.a (mon.0) 61 : cluster [DBG] mgrmap e7: a(active, since 2s) 2026-03-10T10:15:00.014 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout Added host 'vm00' with addr '192.168.123.100' 2026-03-10T10:15:00.014 INFO:teuthology.orchestra.run.vm00.stdout:Deploying unmanaged mon service... 2026-03-10T10:15:00.328 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-10T10:15:00.328 INFO:teuthology.orchestra.run.vm00.stdout:Deploying unmanaged mgr service... 2026-03-10T10:15:00.811 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-10T10:15:00.850 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:00 vm00 bash[20709]: cephadm 2026-03-10T10:14:58.683973+0000 mgr.a (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm00 2026-03-10T10:15:00.851 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:00 vm00 bash[20709]: cephadm 2026-03-10T10:14:58.683973+0000 mgr.a (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm00 2026-03-10T10:15:00.851 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:00 vm00 bash[20709]: audit 2026-03-10T10:14:59.952848+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:00.851 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:00 vm00 bash[20709]: audit 2026-03-10T10:14:59.952848+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:00.851 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:00 vm00 bash[20709]: audit 2026-03-10T10:14:59.953348+0000 mon.a (mon.0) 63 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:00.851 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:00 vm00 bash[20709]: audit 2026-03-10T10:14:59.953348+0000 mon.a (mon.0) 63 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:00.851 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:00 vm00 bash[20709]: audit 2026-03-10T10:15:00.289441+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:00.851 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:00 vm00 bash[20709]: audit 2026-03-10T10:15:00.289441+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:01.496 INFO:teuthology.orchestra.run.vm00.stdout:Enabling the dashboard module... 2026-03-10T10:15:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:01 vm00 bash[20709]: cephadm 2026-03-10T10:14:59.953179+0000 mgr.a (mgr.14118) 16 : cephadm [INF] Added host vm00 2026-03-10T10:15:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:01 vm00 bash[20709]: cephadm 2026-03-10T10:14:59.953179+0000 mgr.a (mgr.14118) 16 : cephadm [INF] Added host vm00 2026-03-10T10:15:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:01 vm00 bash[20709]: audit 2026-03-10T10:15:00.285780+0000 mgr.a (mgr.14118) 17 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:01 vm00 bash[20709]: audit 2026-03-10T10:15:00.285780+0000 mgr.a (mgr.14118) 17 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:01 vm00 bash[20709]: cephadm 2026-03-10T10:15:00.286656+0000 mgr.a (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-10T10:15:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:01 vm00 bash[20709]: cephadm 2026-03-10T10:15:00.286656+0000 mgr.a (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-10T10:15:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:01 vm00 bash[20709]: audit 2026-03-10T10:15:00.772498+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:01 vm00 bash[20709]: audit 2026-03-10T10:15:00.772498+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:01 vm00 bash[20709]: audit 2026-03-10T10:15:01.102548+0000 mon.a (mon.0) 66 : audit [INF] from='client.? 192.168.123.100:0/1210580903' entity='client.admin' 2026-03-10T10:15:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:01 vm00 bash[20709]: audit 2026-03-10T10:15:01.102548+0000 mon.a (mon.0) 66 : audit [INF] from='client.? 192.168.123.100:0/1210580903' entity='client.admin' 2026-03-10T10:15:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:01 vm00 bash[20709]: audit 2026-03-10T10:15:01.414019+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.100:0/3887716780' entity='client.admin' 2026-03-10T10:15:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:01 vm00 bash[20709]: audit 2026-03-10T10:15:01.414019+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.100:0/3887716780' entity='client.admin' 2026-03-10T10:15:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:01 vm00 bash[20709]: audit 2026-03-10T10:15:01.467433+0000 mon.a (mon.0) 68 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:01 vm00 bash[20709]: audit 2026-03-10T10:15:01.467433+0000 mon.a (mon.0) 68 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:02.897 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:02 vm00 bash[20974]: ignoring --setuser ceph since I am not root 2026-03-10T10:15:02.897 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:02 vm00 bash[20974]: ignoring --setgroup ceph since I am not root 2026-03-10T10:15:02.898 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:02 vm00 bash[20974]: debug 2026-03-10T10:15:02.753+0000 7f270b6cf140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T10:15:02.898 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:02 vm00 bash[20974]: debug 2026-03-10T10:15:02.781+0000 7f270b6cf140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T10:15:02.898 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:02 vm00 bash[20709]: audit 2026-03-10T10:15:00.769197+0000 mgr.a (mgr.14118) 19 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:02.898 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:02 vm00 bash[20709]: audit 2026-03-10T10:15:00.769197+0000 mgr.a (mgr.14118) 19 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:02.898 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:02 vm00 bash[20709]: cephadm 2026-03-10T10:15:00.769897+0000 mgr.a (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-10T10:15:02.898 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:02 vm00 bash[20709]: cephadm 2026-03-10T10:15:00.769897+0000 mgr.a (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-10T10:15:02.898 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:02 vm00 bash[20709]: audit 2026-03-10T10:15:01.759591+0000 mon.a (mon.0) 69 : audit [INF] from='client.? 192.168.123.100:0/2840361483' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T10:15:02.898 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:02 vm00 bash[20709]: audit 2026-03-10T10:15:01.759591+0000 mon.a (mon.0) 69 : audit [INF] from='client.? 192.168.123.100:0/2840361483' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T10:15:02.898 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:02 vm00 bash[20709]: audit 2026-03-10T10:15:01.782198+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:02.898 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:02 vm00 bash[20709]: audit 2026-03-10T10:15:01.782198+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:03.168 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:02 vm00 bash[20974]: debug 2026-03-10T10:15:02.893+0000 7f270b6cf140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T10:15:03.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-10T10:15:03.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 8, 2026-03-10T10:15:03.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T10:15:03.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-10T10:15:03.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-10T10:15:03.419 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-10T10:15:03.419 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for the mgr to restart... 2026-03-10T10:15:03.419 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mgr epoch 8... 2026-03-10T10:15:03.470 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:03 vm00 bash[20974]: debug 2026-03-10T10:15:03.197+0000 7f270b6cf140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T10:15:03.799 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:03 vm00 bash[20974]: debug 2026-03-10T10:15:03.713+0000 7f270b6cf140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T10:15:03.799 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:03 vm00 bash[20709]: audit 2026-03-10T10:15:02.635760+0000 mon.a (mon.0) 71 : audit [INF] from='client.? 192.168.123.100:0/2840361483' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T10:15:03.799 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:03 vm00 bash[20709]: audit 2026-03-10T10:15:02.635760+0000 mon.a (mon.0) 71 : audit [INF] from='client.? 192.168.123.100:0/2840361483' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T10:15:03.799 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:03 vm00 bash[20709]: cluster 2026-03-10T10:15:02.638223+0000 mon.a (mon.0) 72 : cluster [DBG] mgrmap e8: a(active, since 6s) 2026-03-10T10:15:03.799 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:03 vm00 bash[20709]: cluster 2026-03-10T10:15:02.638223+0000 mon.a (mon.0) 72 : cluster [DBG] mgrmap e8: a(active, since 6s) 2026-03-10T10:15:03.799 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:03 vm00 bash[20709]: audit 2026-03-10T10:15:03.370528+0000 mon.a (mon.0) 73 : audit [DBG] from='client.? 192.168.123.100:0/1190503004' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T10:15:03.799 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:03 vm00 bash[20709]: audit 2026-03-10T10:15:03.370528+0000 mon.a (mon.0) 73 : audit [DBG] from='client.? 192.168.123.100:0/1190503004' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T10:15:04.087 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:03 vm00 bash[20974]: debug 2026-03-10T10:15:03.793+0000 7f270b6cf140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T10:15:04.087 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:03 vm00 bash[20974]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T10:15:04.087 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:03 vm00 bash[20974]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T10:15:04.087 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:03 vm00 bash[20974]: from numpy import show_config as show_numpy_config 2026-03-10T10:15:04.087 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:03 vm00 bash[20974]: debug 2026-03-10T10:15:03.909+0000 7f270b6cf140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T10:15:04.087 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:04 vm00 bash[20974]: debug 2026-03-10T10:15:04.045+0000 7f270b6cf140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T10:15:04.418 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:04 vm00 bash[20974]: debug 2026-03-10T10:15:04.081+0000 7f270b6cf140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T10:15:04.418 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:04 vm00 bash[20974]: debug 2026-03-10T10:15:04.121+0000 7f270b6cf140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T10:15:04.418 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:04 vm00 bash[20974]: debug 2026-03-10T10:15:04.161+0000 7f270b6cf140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T10:15:04.418 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:04 vm00 bash[20974]: debug 2026-03-10T10:15:04.209+0000 7f270b6cf140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T10:15:04.901 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:04 vm00 bash[20974]: debug 2026-03-10T10:15:04.617+0000 7f270b6cf140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T10:15:04.901 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:04 vm00 bash[20974]: debug 2026-03-10T10:15:04.653+0000 7f270b6cf140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T10:15:04.901 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:04 vm00 bash[20974]: debug 2026-03-10T10:15:04.685+0000 7f270b6cf140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T10:15:04.901 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:04 vm00 bash[20974]: debug 2026-03-10T10:15:04.821+0000 7f270b6cf140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T10:15:04.901 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:04 vm00 bash[20974]: debug 2026-03-10T10:15:04.861+0000 7f270b6cf140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T10:15:05.168 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:04 vm00 bash[20974]: debug 2026-03-10T10:15:04.897+0000 7f270b6cf140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T10:15:05.168 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20974]: debug 2026-03-10T10:15:04.997+0000 7f270b6cf140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T10:15:05.168 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20974]: debug 2026-03-10T10:15:05.141+0000 7f270b6cf140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T10:15:05.668 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20974]: debug 2026-03-10T10:15:05.293+0000 7f270b6cf140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T10:15:05.668 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20974]: debug 2026-03-10T10:15:05.325+0000 7f270b6cf140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T10:15:05.668 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20974]: debug 2026-03-10T10:15:05.365+0000 7f270b6cf140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T10:15:05.668 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20974]: debug 2026-03-10T10:15:05.505+0000 7f270b6cf140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T10:15:06.168 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20974]: debug 2026-03-10T10:15:05.749+0000 7f270b6cf140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T10:15:06.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: cluster 2026-03-10T10:15:05.753510+0000 mon.a (mon.0) 74 : cluster [INF] Active manager daemon a restarted 2026-03-10T10:15:06.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: cluster 2026-03-10T10:15:05.753510+0000 mon.a (mon.0) 74 : cluster [INF] Active manager daemon a restarted 2026-03-10T10:15:06.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: cluster 2026-03-10T10:15:05.753731+0000 mon.a (mon.0) 75 : cluster [INF] Activating manager daemon a 2026-03-10T10:15:06.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: cluster 2026-03-10T10:15:05.753731+0000 mon.a (mon.0) 75 : cluster [INF] Activating manager daemon a 2026-03-10T10:15:06.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: cluster 2026-03-10T10:15:05.759504+0000 mon.a (mon.0) 76 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-10T10:15:06.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: cluster 2026-03-10T10:15:05.759504+0000 mon.a (mon.0) 76 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-10T10:15:06.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: cluster 2026-03-10T10:15:05.759678+0000 mon.a (mon.0) 77 : cluster [DBG] mgrmap e9: a(active, starting, since 0.00603039s) 2026-03-10T10:15:06.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: cluster 2026-03-10T10:15:05.759678+0000 mon.a (mon.0) 77 : cluster [DBG] mgrmap e9: a(active, starting, since 0.00603039s) 2026-03-10T10:15:06.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.761353+0000 mon.a (mon.0) 78 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:06.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.761353+0000 mon.a (mon.0) 78 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:06.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.761760+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:15:06.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.761760+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:15:06.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.762588+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:15:06.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.762588+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:15:06.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.762745+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:15:06.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.762745+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:15:06.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.762934+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:15:06.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.762934+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:15:06.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: cluster 2026-03-10T10:15:05.768935+0000 mon.a (mon.0) 83 : cluster [INF] Manager daemon a is now available 2026-03-10T10:15:06.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: cluster 2026-03-10T10:15:05.768935+0000 mon.a (mon.0) 83 : cluster [INF] Manager daemon a is now available 2026-03-10T10:15:06.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.781428+0000 mon.a (mon.0) 84 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:06.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.781428+0000 mon.a (mon.0) 84 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:06.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.791322+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:06.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.791322+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:06.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.796258+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:15:06.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.796258+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:15:06.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.803146+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:15:06.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:05 vm00 bash[20709]: audit 2026-03-10T10:15:05.803146+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:15:06.811 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-10T10:15:06.811 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 10, 2026-03-10T10:15:06.811 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-10T10:15:06.811 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-10T10:15:06.811 INFO:teuthology.orchestra.run.vm00.stdout:mgr epoch 8 is available 2026-03-10T10:15:06.811 INFO:teuthology.orchestra.run.vm00.stdout:Generating a dashboard self-signed certificate... 2026-03-10T10:15:07.092 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-10T10:15:07.092 INFO:teuthology.orchestra.run.vm00.stdout:Creating initial admin user... 2026-03-10T10:15:07.482 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$.d.LClkBVxYAmLRQI2nF4OPyYknl8Wxye.evrHaq1Xgkro9dRMVaq", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1773137707, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-10T10:15:07.482 INFO:teuthology.orchestra.run.vm00.stdout:Fetching dashboard port number... 2026-03-10T10:15:07.760 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 8443 2026-03-10T10:15:07.760 INFO:teuthology.orchestra.run.vm00.stdout:firewalld does not appear to be present 2026-03-10T10:15:07.760 INFO:teuthology.orchestra.run.vm00.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-10T10:15:07.760 INFO:teuthology.orchestra.run.vm00.stdout:Ceph Dashboard is now available at: 2026-03-10T10:15:07.760 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:15:07.760 INFO:teuthology.orchestra.run.vm00.stdout: URL: https://vm00.local:8443/ 2026-03-10T10:15:07.760 INFO:teuthology.orchestra.run.vm00.stdout: User: admin 2026-03-10T10:15:07.760 INFO:teuthology.orchestra.run.vm00.stdout: Password: odl4j1s8ac 2026-03-10T10:15:07.760 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:15:07.760 INFO:teuthology.orchestra.run.vm00.stdout:Saving cluster configuration to /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config directory 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: cluster 2026-03-10T10:15:06.762888+0000 mon.a (mon.0) 88 : cluster [DBG] mgrmap e10: a(active, since 1.00924s) 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: cluster 2026-03-10T10:15:06.762888+0000 mon.a (mon.0) 88 : cluster [DBG] mgrmap e10: a(active, since 1.00924s) 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: cephadm 2026-03-10T10:15:06.855107+0000 mgr.a (mgr.14150) 3 : cephadm [INF] [10/Mar/2026:10:15:06] ENGINE Bus STARTING 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: cephadm 2026-03-10T10:15:06.855107+0000 mgr.a (mgr.14150) 3 : cephadm [INF] [10/Mar/2026:10:15:06] ENGINE Bus STARTING 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: cephadm 2026-03-10T10:15:06.956306+0000 mgr.a (mgr.14150) 4 : cephadm [INF] [10/Mar/2026:10:15:06] ENGINE Serving on http://192.168.123.100:8765 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: cephadm 2026-03-10T10:15:06.956306+0000 mgr.a (mgr.14150) 4 : cephadm [INF] [10/Mar/2026:10:15:06] ENGINE Serving on http://192.168.123.100:8765 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: audit 2026-03-10T10:15:07.026151+0000 mgr.a (mgr.14150) 5 : audit [DBG] from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: audit 2026-03-10T10:15:07.026151+0000 mgr.a (mgr.14150) 5 : audit [DBG] from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: audit 2026-03-10T10:15:07.054291+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: audit 2026-03-10T10:15:07.054291+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: audit 2026-03-10T10:15:07.056813+0000 mon.a (mon.0) 90 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: audit 2026-03-10T10:15:07.056813+0000 mon.a (mon.0) 90 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: cephadm 2026-03-10T10:15:07.065557+0000 mgr.a (mgr.14150) 6 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Serving on https://192.168.123.100:7150 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: cephadm 2026-03-10T10:15:07.065557+0000 mgr.a (mgr.14150) 6 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Serving on https://192.168.123.100:7150 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: cephadm 2026-03-10T10:15:07.065605+0000 mgr.a (mgr.14150) 7 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Bus STARTED 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: cephadm 2026-03-10T10:15:07.065605+0000 mgr.a (mgr.14150) 7 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Bus STARTED 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: cephadm 2026-03-10T10:15:07.065900+0000 mgr.a (mgr.14150) 8 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Client ('192.168.123.100', 40024) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: cephadm 2026-03-10T10:15:07.065900+0000 mgr.a (mgr.14150) 8 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Client ('192.168.123.100', 40024) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: audit 2026-03-10T10:15:07.297971+0000 mgr.a (mgr.14150) 9 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: audit 2026-03-10T10:15:07.297971+0000 mgr.a (mgr.14150) 9 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: audit 2026-03-10T10:15:07.447019+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: audit 2026-03-10T10:15:07.447019+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: audit 2026-03-10T10:15:07.721638+0000 mon.a (mon.0) 92 : audit [DBG] from='client.? 192.168.123.100:0/1740426705' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T10:15:08.060 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:07 vm00 bash[20709]: audit 2026-03-10T10:15:07.721638+0000 mon.a (mon.0) 92 : audit [DBG] from='client.? 192.168.123.100:0/1740426705' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout:Or, if you are only running a single cluster on this host: 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout: ceph telemetry on 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout:For more information see: 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:15:08.403 INFO:teuthology.orchestra.run.vm00.stdout:Bootstrap complete. 2026-03-10T10:15:08.422 INFO:tasks.cephadm:Fetching config... 2026-03-10T10:15:08.422 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-10T10:15:08.422 DEBUG:teuthology.orchestra.run.vm00:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-10T10:15:08.425 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-10T10:15:08.425 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-10T10:15:08.425 DEBUG:teuthology.orchestra.run.vm00:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-10T10:15:08.470 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-10T10:15:08.470 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-10T10:15:08.470 DEBUG:teuthology.orchestra.run.vm00:> sudo dd if=/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/keyring of=/dev/stdout 2026-03-10T10:15:08.519 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-10T10:15:08.519 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-10T10:15:08.519 DEBUG:teuthology.orchestra.run.vm00:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-10T10:15:08.562 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-10T10:15:08.562 DEBUG:teuthology.orchestra.run.vm00:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIKXzt4Lk4e3wujrSBUxsCaFaFvtK1ABc232NZhzsYY1G ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T10:15:08.615 INFO:teuthology.orchestra.run.vm00.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIKXzt4Lk4e3wujrSBUxsCaFaFvtK1ABc232NZhzsYY1G ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:08.622 DEBUG:teuthology.orchestra.run.vm03:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIKXzt4Lk4e3wujrSBUxsCaFaFvtK1ABc232NZhzsYY1G ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T10:15:08.635 INFO:teuthology.orchestra.run.vm03.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIKXzt4Lk4e3wujrSBUxsCaFaFvtK1ABc232NZhzsYY1G ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:08.640 DEBUG:teuthology.orchestra.run.vm09:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIKXzt4Lk4e3wujrSBUxsCaFaFvtK1ABc232NZhzsYY1G ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T10:15:08.652 INFO:teuthology.orchestra.run.vm09.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIKXzt4Lk4e3wujrSBUxsCaFaFvtK1ABc232NZhzsYY1G ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:08.657 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-10T10:15:09.050 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:09 vm00 bash[20709]: audit 2026-03-10T10:15:08.048363+0000 mon.a (mon.0) 93 : audit [INF] from='client.? 192.168.123.100:0/2918787650' entity='client.admin' 2026-03-10T10:15:09.050 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:09 vm00 bash[20709]: audit 2026-03-10T10:15:08.048363+0000 mon.a (mon.0) 93 : audit [INF] from='client.? 192.168.123.100:0/2918787650' entity='client.admin' 2026-03-10T10:15:09.050 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:09 vm00 bash[20709]: cluster 2026-03-10T10:15:08.451766+0000 mon.a (mon.0) 94 : cluster [DBG] mgrmap e11: a(active, since 2s) 2026-03-10T10:15:09.050 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:09 vm00 bash[20709]: cluster 2026-03-10T10:15:08.451766+0000 mon.a (mon.0) 94 : cluster [DBG] mgrmap e11: a(active, since 2s) 2026-03-10T10:15:11.987 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:11 vm00 bash[20709]: audit 2026-03-10T10:15:10.693049+0000 mon.a (mon.0) 95 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:11.987 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:11 vm00 bash[20709]: audit 2026-03-10T10:15:10.693049+0000 mon.a (mon.0) 95 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:11.987 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:11 vm00 bash[20709]: audit 2026-03-10T10:15:10.693786+0000 mon.a (mon.0) 96 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:11.987 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:11 vm00 bash[20709]: audit 2026-03-10T10:15:10.693786+0000 mon.a (mon.0) 96 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:11.987 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:11 vm00 bash[20709]: audit 2026-03-10T10:15:10.696540+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:11.987 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:11 vm00 bash[20709]: audit 2026-03-10T10:15:10.696540+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:11.988 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:11 vm00 bash[20709]: audit 2026-03-10T10:15:10.697444+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]: dispatch 2026-03-10T10:15:11.988 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:11 vm00 bash[20709]: audit 2026-03-10T10:15:10.697444+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]: dispatch 2026-03-10T10:15:11.988 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:11 vm00 bash[20709]: audit 2026-03-10T10:15:10.698653+0000 mon.a (mon.0) 99 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]': finished 2026-03-10T10:15:11.988 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:11 vm00 bash[20709]: audit 2026-03-10T10:15:10.698653+0000 mon.a (mon.0) 99 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]': finished 2026-03-10T10:15:11.988 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:11 vm00 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:15:11.988 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:11 vm00 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:15:12.277 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:11 vm00 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:15:12.277 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:11 vm00 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:15:12.559 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: cephadm 2026-03-10T10:15:11.443247+0000 mgr.a (mgr.14150) 10 : cephadm [INF] Deploying daemon agent.vm00 on vm00 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: cephadm 2026-03-10T10:15:11.443247+0000 mgr.a (mgr.14150) 10 : cephadm [INF] Deploying daemon agent.vm00 on vm00 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.072226+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.072226+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.074786+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.074786+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.076813+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.076813+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.082261+0000 mon.a (mon.0) 103 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.082261+0000 mon.a (mon.0) 103 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.085755+0000 mon.a (mon.0) 104 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.085755+0000 mon.a (mon.0) 104 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.092222+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.092222+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.094717+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.094717+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.096816+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.096816+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.099935+0000 mon.a (mon.0) 108 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.099935+0000 mon.a (mon.0) 108 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.106233+0000 mon.a (mon.0) 109 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.106233+0000 mon.a (mon.0) 109 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.109685+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.109685+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.349600+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.826 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:12 vm00 bash[20709]: audit 2026-03-10T10:15:12.349600+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:12.858 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-10T10:15:12.858 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-10T10:15:13.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:13 vm00 bash[20709]: cluster 2026-03-10T10:15:12.702263+0000 mon.a (mon.0) 112 : cluster [DBG] mgrmap e12: a(active, since 6s) 2026-03-10T10:15:13.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:13 vm00 bash[20709]: cluster 2026-03-10T10:15:12.702263+0000 mon.a (mon.0) 112 : cluster [DBG] mgrmap e12: a(active, since 6s) 2026-03-10T10:15:13.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:13 vm00 bash[20709]: audit 2026-03-10T10:15:12.797145+0000 mon.a (mon.0) 113 : audit [INF] from='client.? 192.168.123.100:0/3135075601' entity='client.admin' 2026-03-10T10:15:13.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:13 vm00 bash[20709]: audit 2026-03-10T10:15:12.797145+0000 mon.a (mon.0) 113 : audit [INF] from='client.? 192.168.123.100:0/3135075601' entity='client.admin' 2026-03-10T10:15:13.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:13 vm00 bash[20709]: audit 2026-03-10T10:15:12.802223+0000 mon.a (mon.0) 114 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:13.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:13 vm00 bash[20709]: audit 2026-03-10T10:15:12.802223+0000 mon.a (mon.0) 114 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:13.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:13 vm00 bash[20709]: audit 2026-03-10T10:15:12.807545+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:13.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:13 vm00 bash[20709]: audit 2026-03-10T10:15:12.807545+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:16.570 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:15:16.855 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:16 vm00 bash[20709]: audit 2026-03-10T10:15:15.793178+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:16.856 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:16 vm00 bash[20709]: audit 2026-03-10T10:15:15.793178+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:16.936 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm03 2026-03-10T10:15:16.937 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-10T10:15:16.937 DEBUG:teuthology.orchestra.run.vm03:> dd of=/etc/ceph/ceph.conf 2026-03-10T10:15:16.940 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-10T10:15:16.940 DEBUG:teuthology.orchestra.run.vm03:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:16.985 INFO:tasks.cephadm:Adding host vm03 to orchestrator... 2026-03-10T10:15:16.985 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph orch host add vm03 2026-03-10T10:15:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.851569+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.851569+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.853957+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.853957+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.858528+0000 mgr.a (mgr.14150) 11 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.858528+0000 mgr.a (mgr.14150) 11 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.860037+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.860037+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.860448+0000 mon.a (mon.0) 120 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.860448+0000 mon.a (mon.0) 120 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.861276+0000 mon.a (mon.0) 121 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:18.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.861276+0000 mon.a (mon.0) 121 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:18.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.861627+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:18.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.861627+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:18.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: cephadm 2026-03-10T10:15:16.862073+0000 mgr.a (mgr.14150) 12 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:15:18.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: cephadm 2026-03-10T10:15:16.862073+0000 mgr.a (mgr.14150) 12 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:15:18.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: cephadm 2026-03-10T10:15:16.892600+0000 mgr.a (mgr.14150) 13 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:18.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: cephadm 2026-03-10T10:15:16.892600+0000 mgr.a (mgr.14150) 13 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:18.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: cephadm 2026-03-10T10:15:16.932604+0000 mgr.a (mgr.14150) 14 : cephadm [INF] Updating vm00:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:18.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: cephadm 2026-03-10T10:15:16.932604+0000 mgr.a (mgr.14150) 14 : cephadm [INF] Updating vm00:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:18.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: cephadm 2026-03-10T10:15:16.961644+0000 mgr.a (mgr.14150) 15 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:18.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: cephadm 2026-03-10T10:15:16.961644+0000 mgr.a (mgr.14150) 15 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:18.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.996487+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:18.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.996487+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:18.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.999145+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:18.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:17 vm00 bash[20709]: audit 2026-03-10T10:15:16.999145+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:22.334 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:15:23.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:22 vm00 bash[20709]: audit 2026-03-10T10:15:21.883952+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:23.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:22 vm00 bash[20709]: audit 2026-03-10T10:15:21.883952+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:23.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:22 vm00 bash[20709]: audit 2026-03-10T10:15:21.886590+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:23.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:22 vm00 bash[20709]: audit 2026-03-10T10:15:21.886590+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:23.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:22 vm00 bash[20709]: audit 2026-03-10T10:15:22.565126+0000 mgr.a (mgr.14150) 16 : audit [DBG] from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm03", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:23.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:22 vm00 bash[20709]: audit 2026-03-10T10:15:22.565126+0000 mgr.a (mgr.14150) 16 : audit [DBG] from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm03", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:24.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:23 vm00 bash[20709]: cephadm 2026-03-10T10:15:23.090899+0000 mgr.a (mgr.14150) 17 : cephadm [INF] Deploying cephadm binary to vm03 2026-03-10T10:15:24.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:23 vm00 bash[20709]: cephadm 2026-03-10T10:15:23.090899+0000 mgr.a (mgr.14150) 17 : cephadm [INF] Deploying cephadm binary to vm03 2026-03-10T10:15:24.286 INFO:teuthology.orchestra.run.vm00.stdout:Added host 'vm03' with addr '192.168.123.103' 2026-03-10T10:15:24.352 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph orch host ls --format=json 2026-03-10T10:15:25.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:25 vm00 bash[20709]: audit 2026-03-10T10:15:24.284271+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:25.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:25 vm00 bash[20709]: audit 2026-03-10T10:15:24.284271+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:25.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:25 vm00 bash[20709]: cephadm 2026-03-10T10:15:24.284686+0000 mgr.a (mgr.14150) 18 : cephadm [INF] Added host vm03 2026-03-10T10:15:25.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:25 vm00 bash[20709]: cephadm 2026-03-10T10:15:24.284686+0000 mgr.a (mgr.14150) 18 : cephadm [INF] Added host vm03 2026-03-10T10:15:25.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:25 vm00 bash[20709]: audit 2026-03-10T10:15:24.285038+0000 mon.a (mon.0) 128 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:25.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:25 vm00 bash[20709]: audit 2026-03-10T10:15:24.285038+0000 mon.a (mon.0) 128 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:25.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:25 vm00 bash[20709]: audit 2026-03-10T10:15:24.557335+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:25.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:25 vm00 bash[20709]: audit 2026-03-10T10:15:24.557335+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:27.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: cluster 2026-03-10T10:15:25.763446+0000 mgr.a (mgr.14150) 19 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:27.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: cluster 2026-03-10T10:15:25.763446+0000 mgr.a (mgr.14150) 19 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:27.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.815094+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:27.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.815094+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:27.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.817730+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:27.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.817730+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:27.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.818256+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:27.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.818256+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:27.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.818902+0000 mon.a (mon.0) 133 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:27.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.818902+0000 mon.a (mon.0) 133 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.819301+0000 mon.a (mon.0) 134 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.819301+0000 mon.a (mon.0) 134 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: cephadm 2026-03-10T10:15:25.819855+0000 mgr.a (mgr.14150) 20 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: cephadm 2026-03-10T10:15:25.819855+0000 mgr.a (mgr.14150) 20 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: cephadm 2026-03-10T10:15:25.848609+0000 mgr.a (mgr.14150) 21 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: cephadm 2026-03-10T10:15:25.848609+0000 mgr.a (mgr.14150) 21 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: cephadm 2026-03-10T10:15:25.877234+0000 mgr.a (mgr.14150) 22 : cephadm [INF] Updating vm03:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: cephadm 2026-03-10T10:15:25.877234+0000 mgr.a (mgr.14150) 22 : cephadm [INF] Updating vm03:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: cephadm 2026-03-10T10:15:25.911845+0000 mgr.a (mgr.14150) 23 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: cephadm 2026-03-10T10:15:25.911845+0000 mgr.a (mgr.14150) 23 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.944959+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.944959+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.947071+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.947071+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.947867+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]: dispatch 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.947867+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]: dispatch 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.948841+0000 mon.a (mon.0) 138 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]': finished 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: audit 2026-03-10T10:15:25.948841+0000 mon.a (mon.0) 138 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]': finished 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: cephadm 2026-03-10T10:15:26.326527+0000 mgr.a (mgr.14150) 24 : cephadm [INF] Deploying daemon agent.vm03 on vm03 2026-03-10T10:15:27.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:26 vm00 bash[20709]: cephadm 2026-03-10T10:15:26.326527+0000 mgr.a (mgr.14150) 24 : cephadm [INF] Deploying daemon agent.vm03 on vm03 2026-03-10T10:15:28.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:26.971739+0000 mon.a (mon.0) 139 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:28.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:26.971739+0000 mon.a (mon.0) 139 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:28.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:26.973943+0000 mon.a (mon.0) 140 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:28.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:26.973943+0000 mon.a (mon.0) 140 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:28.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:26.976613+0000 mon.a (mon.0) 141 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:26.976613+0000 mon.a (mon.0) 141 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:26.986160+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:26.986160+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:26.987509+0000 mon.a (mon.0) 143 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:26.987509+0000 mon.a (mon.0) 143 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:26.988036+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:26.988036+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:26.991688+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:26.991688+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:27.003194+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:27.003194+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:27.004316+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:27.004316+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:27.004722+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:27.004722+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:27.007998+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:27.007998+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:27.231900+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:27 vm00 bash[20709]: audit 2026-03-10T10:15:27.231900+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:28.959 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:15:29.208 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:15:29.208 INFO:teuthology.orchestra.run.vm00.stdout:[{"addr": "192.168.123.100", "hostname": "vm00", "labels": [], "status": ""}, {"addr": "192.168.123.103", "hostname": "vm03", "labels": [], "status": ""}] 2026-03-10T10:15:29.253 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm09 2026-03-10T10:15:29.253 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T10:15:29.253 DEBUG:teuthology.orchestra.run.vm09:> dd of=/etc/ceph/ceph.conf 2026-03-10T10:15:29.256 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T10:15:29.256 DEBUG:teuthology.orchestra.run.vm09:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:29.301 INFO:tasks.cephadm:Adding host vm09 to orchestrator... 2026-03-10T10:15:29.301 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph orch host add vm09 2026-03-10T10:15:29.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:29 vm00 bash[20709]: cluster 2026-03-10T10:15:27.763612+0000 mgr.a (mgr.14150) 25 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:29.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:29 vm00 bash[20709]: cluster 2026-03-10T10:15:27.763612+0000 mgr.a (mgr.14150) 25 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:29.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:29 vm00 bash[20709]: audit 2026-03-10T10:15:28.276683+0000 mon.a (mon.0) 151 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:29.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:29 vm00 bash[20709]: audit 2026-03-10T10:15:28.276683+0000 mon.a (mon.0) 151 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:29.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:29 vm00 bash[20709]: audit 2026-03-10T10:15:28.279096+0000 mon.a (mon.0) 152 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:29.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:29 vm00 bash[20709]: audit 2026-03-10T10:15:28.279096+0000 mon.a (mon.0) 152 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:30.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:30 vm00 bash[20709]: audit 2026-03-10T10:15:29.206326+0000 mgr.a (mgr.14150) 26 : audit [DBG] from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:15:30.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:30 vm00 bash[20709]: audit 2026-03-10T10:15:29.206326+0000 mgr.a (mgr.14150) 26 : audit [DBG] from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:15:31.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:31 vm00 bash[20709]: cluster 2026-03-10T10:15:29.763764+0000 mgr.a (mgr.14150) 27 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:31.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:31 vm00 bash[20709]: cluster 2026-03-10T10:15:29.763764+0000 mgr.a (mgr.14150) 27 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:31.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:31 vm00 bash[20709]: audit 2026-03-10T10:15:30.796868+0000 mon.a (mon.0) 153 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:31.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:31 vm00 bash[20709]: audit 2026-03-10T10:15:30.796868+0000 mon.a (mon.0) 153 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:32.968 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:15:33.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:33 vm00 bash[20709]: cluster 2026-03-10T10:15:31.763932+0000 mgr.a (mgr.14150) 28 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:33.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:33 vm00 bash[20709]: cluster 2026-03-10T10:15:31.763932+0000 mgr.a (mgr.14150) 28 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:34.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:34 vm00 bash[20709]: audit 2026-03-10T10:15:33.256409+0000 mgr.a (mgr.14150) 29 : audit [DBG] from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:34.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:34 vm00 bash[20709]: audit 2026-03-10T10:15:33.256409+0000 mgr.a (mgr.14150) 29 : audit [DBG] from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:34.957 INFO:teuthology.orchestra.run.vm00.stdout:Added host 'vm09' with addr '192.168.123.109' 2026-03-10T10:15:35.027 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph orch host ls --format=json 2026-03-10T10:15:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:35 vm00 bash[20709]: cephadm 2026-03-10T10:15:33.762886+0000 mgr.a (mgr.14150) 30 : cephadm [INF] Deploying cephadm binary to vm09 2026-03-10T10:15:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:35 vm00 bash[20709]: cephadm 2026-03-10T10:15:33.762886+0000 mgr.a (mgr.14150) 30 : cephadm [INF] Deploying cephadm binary to vm09 2026-03-10T10:15:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:35 vm00 bash[20709]: cluster 2026-03-10T10:15:33.764093+0000 mgr.a (mgr.14150) 31 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:35 vm00 bash[20709]: cluster 2026-03-10T10:15:33.764093+0000 mgr.a (mgr.14150) 31 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:35 vm00 bash[20709]: audit 2026-03-10T10:15:34.955366+0000 mon.a (mon.0) 154 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:35 vm00 bash[20709]: audit 2026-03-10T10:15:34.955366+0000 mon.a (mon.0) 154 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:35 vm00 bash[20709]: audit 2026-03-10T10:15:34.956369+0000 mon.a (mon.0) 155 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:35 vm00 bash[20709]: audit 2026-03-10T10:15:34.956369+0000 mon.a (mon.0) 155 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:35 vm00 bash[20709]: audit 2026-03-10T10:15:35.073114+0000 mon.a (mon.0) 156 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:35 vm00 bash[20709]: audit 2026-03-10T10:15:35.073114+0000 mon.a (mon.0) 156 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:35 vm00 bash[20709]: audit 2026-03-10T10:15:35.076842+0000 mon.a (mon.0) 157 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:35 vm00 bash[20709]: audit 2026-03-10T10:15:35.076842+0000 mon.a (mon.0) 157 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:35 vm00 bash[20709]: audit 2026-03-10T10:15:35.090473+0000 mon.a (mon.0) 158 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:35 vm00 bash[20709]: audit 2026-03-10T10:15:35.090473+0000 mon.a (mon.0) 158 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:35 vm00 bash[20709]: audit 2026-03-10T10:15:35.240644+0000 mon.a (mon.0) 159 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:35 vm00 bash[20709]: audit 2026-03-10T10:15:35.240644+0000 mon.a (mon.0) 159 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:36.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:36 vm00 bash[20709]: cephadm 2026-03-10T10:15:34.956152+0000 mgr.a (mgr.14150) 32 : cephadm [INF] Added host vm09 2026-03-10T10:15:36.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:36 vm00 bash[20709]: cephadm 2026-03-10T10:15:34.956152+0000 mgr.a (mgr.14150) 32 : cephadm [INF] Added host vm09 2026-03-10T10:15:37.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: cluster 2026-03-10T10:15:35.764276+0000 mgr.a (mgr.14150) 33 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:37.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: cluster 2026-03-10T10:15:35.764276+0000 mgr.a (mgr.14150) 33 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:37.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.498998+0000 mon.a (mon.0) 160 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:37.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.498998+0000 mon.a (mon.0) 160 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:37.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.500913+0000 mon.a (mon.0) 161 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:37.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.500913+0000 mon.a (mon.0) 161 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:37.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.501352+0000 mon.a (mon.0) 162 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:37.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.501352+0000 mon.a (mon.0) 162 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:37.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.501943+0000 mon.a (mon.0) 163 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:37.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.501943+0000 mon.a (mon.0) 163 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:37.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.502308+0000 mon.a (mon.0) 164 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:37.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.502308+0000 mon.a (mon.0) 164 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:37.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: cephadm 2026-03-10T10:15:36.502878+0000 mgr.a (mgr.14150) 34 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:15:37.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: cephadm 2026-03-10T10:15:36.502878+0000 mgr.a (mgr.14150) 34 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:15:37.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: cephadm 2026-03-10T10:15:36.529054+0000 mgr.a (mgr.14150) 35 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:37.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: cephadm 2026-03-10T10:15:36.529054+0000 mgr.a (mgr.14150) 35 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:37.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: cephadm 2026-03-10T10:15:36.553286+0000 mgr.a (mgr.14150) 36 : cephadm [INF] Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:37.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: cephadm 2026-03-10T10:15:36.553286+0000 mgr.a (mgr.14150) 36 : cephadm [INF] Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:37.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: cephadm 2026-03-10T10:15:36.577465+0000 mgr.a (mgr.14150) 37 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:37.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: cephadm 2026-03-10T10:15:36.577465+0000 mgr.a (mgr.14150) 37 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:37.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.604392+0000 mon.a (mon.0) 165 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:37.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.604392+0000 mon.a (mon.0) 165 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:37.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.606964+0000 mon.a (mon.0) 166 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:37.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.606964+0000 mon.a (mon.0) 166 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:37.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.607844+0000 mon.a (mon.0) 167 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []}]: dispatch 2026-03-10T10:15:37.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.607844+0000 mon.a (mon.0) 167 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []}]: dispatch 2026-03-10T10:15:37.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.608867+0000 mon.a (mon.0) 168 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []}]': finished 2026-03-10T10:15:37.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:37 vm00 bash[20709]: audit 2026-03-10T10:15:36.608867+0000 mon.a (mon.0) 168 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []}]': finished 2026-03-10T10:15:38.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: cephadm 2026-03-10T10:15:37.024492+0000 mgr.a (mgr.14150) 38 : cephadm [INF] Deploying daemon agent.vm09 on vm09 2026-03-10T10:15:38.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: cephadm 2026-03-10T10:15:37.024492+0000 mgr.a (mgr.14150) 38 : cephadm [INF] Deploying daemon agent.vm09 on vm09 2026-03-10T10:15:38.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.640248+0000 mon.a (mon.0) 169 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:38.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.640248+0000 mon.a (mon.0) 169 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:38.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.643021+0000 mon.a (mon.0) 170 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:38.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.643021+0000 mon.a (mon.0) 170 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:38.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.645207+0000 mon.a (mon.0) 171 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:38.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.645207+0000 mon.a (mon.0) 171 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:38.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.660173+0000 mon.a (mon.0) 172 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:38.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.660173+0000 mon.a (mon.0) 172 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:38.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.661491+0000 mon.a (mon.0) 173 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:38.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.661491+0000 mon.a (mon.0) 173 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:38.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.661925+0000 mon.a (mon.0) 174 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:38.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.661925+0000 mon.a (mon.0) 174 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:38.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.665227+0000 mon.a (mon.0) 175 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:38.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.665227+0000 mon.a (mon.0) 175 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:38.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.678573+0000 mon.a (mon.0) 176 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:38.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.678573+0000 mon.a (mon.0) 176 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:38.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.679711+0000 mon.a (mon.0) 177 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:38.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.679711+0000 mon.a (mon.0) 177 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:38.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.680085+0000 mon.a (mon.0) 178 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:38.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.680085+0000 mon.a (mon.0) 178 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:38.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.682903+0000 mon.a (mon.0) 179 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:38.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.682903+0000 mon.a (mon.0) 179 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:38.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.908612+0000 mon.a (mon.0) 180 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:38.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:37.908612+0000 mon.a (mon.0) 180 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:38.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:38.623828+0000 mon.a (mon.0) 181 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:38.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:38.623828+0000 mon.a (mon.0) 181 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:38.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:38.626247+0000 mon.a (mon.0) 182 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:38.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:38 vm00 bash[20709]: audit 2026-03-10T10:15:38.626247+0000 mon.a (mon.0) 182 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:38.983 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:15:39.547 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:15:39.547 INFO:teuthology.orchestra.run.vm00.stdout:[{"addr": "192.168.123.100", "hostname": "vm00", "labels": [], "status": ""}, {"addr": "192.168.123.103", "hostname": "vm03", "labels": [], "status": ""}, {"addr": "192.168.123.109", "hostname": "vm09", "labels": [], "status": ""}] 2026-03-10T10:15:39.594 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-10T10:15:39.594 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph osd crush tunables default 2026-03-10T10:15:39.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:39 vm00 bash[20709]: cluster 2026-03-10T10:15:37.764466+0000 mgr.a (mgr.14150) 39 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:39.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:39 vm00 bash[20709]: cluster 2026-03-10T10:15:37.764466+0000 mgr.a (mgr.14150) 39 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:39.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:39 vm00 bash[20709]: audit 2026-03-10T10:15:38.958192+0000 mon.a (mon.0) 183 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:39.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:39 vm00 bash[20709]: audit 2026-03-10T10:15:38.958192+0000 mon.a (mon.0) 183 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:39.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:39 vm00 bash[20709]: audit 2026-03-10T10:15:38.960842+0000 mon.a (mon.0) 184 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:39.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:39 vm00 bash[20709]: audit 2026-03-10T10:15:38.960842+0000 mon.a (mon.0) 184 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:41.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:40 vm00 bash[20709]: audit 2026-03-10T10:15:39.545387+0000 mgr.a (mgr.14150) 40 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:15:41.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:40 vm00 bash[20709]: audit 2026-03-10T10:15:39.545387+0000 mgr.a (mgr.14150) 40 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:15:42.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:41 vm00 bash[20709]: cluster 2026-03-10T10:15:39.764623+0000 mgr.a (mgr.14150) 41 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:42.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:41 vm00 bash[20709]: cluster 2026-03-10T10:15:39.764623+0000 mgr.a (mgr.14150) 41 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:42.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:41 vm00 bash[20709]: audit 2026-03-10T10:15:40.800636+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:42.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:41 vm00 bash[20709]: audit 2026-03-10T10:15:40.800636+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:44.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:43 vm00 bash[20709]: cluster 2026-03-10T10:15:41.764784+0000 mgr.a (mgr.14150) 42 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:44.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:43 vm00 bash[20709]: cluster 2026-03-10T10:15:41.764784+0000 mgr.a (mgr.14150) 42 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:44.202 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:15:44.700 INFO:teuthology.orchestra.run.vm00.stderr:adjusted tunables profile to default 2026-03-10T10:15:44.755 INFO:tasks.cephadm:Adding mon.a on vm00 2026-03-10T10:15:44.755 INFO:tasks.cephadm:Adding mon.b on vm03 2026-03-10T10:15:44.755 INFO:tasks.cephadm:Adding mon.c on vm09 2026-03-10T10:15:44.755 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph orch apply mon '3;vm00:192.168.123.100=a;vm03:192.168.123.103=b;vm09:192.168.123.109=c' 2026-03-10T10:15:45.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:44 vm00 bash[20709]: audit 2026-03-10T10:15:44.455176+0000 mon.a (mon.0) 186 : audit [INF] from='client.? 192.168.123.100:0/3867082144' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T10:15:45.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:44 vm00 bash[20709]: audit 2026-03-10T10:15:44.455176+0000 mon.a (mon.0) 186 : audit [INF] from='client.? 192.168.123.100:0/3867082144' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T10:15:45.860 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:46.123 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled mon update... 2026-03-10T10:15:46.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:45 vm00 bash[20709]: cluster 2026-03-10T10:15:43.764968+0000 mgr.a (mgr.14150) 43 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:46.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:45 vm00 bash[20709]: cluster 2026-03-10T10:15:43.764968+0000 mgr.a (mgr.14150) 43 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:46.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:45 vm00 bash[20709]: audit 2026-03-10T10:15:44.699288+0000 mon.a (mon.0) 187 : audit [INF] from='client.? 192.168.123.100:0/3867082144' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T10:15:46.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:45 vm00 bash[20709]: audit 2026-03-10T10:15:44.699288+0000 mon.a (mon.0) 187 : audit [INF] from='client.? 192.168.123.100:0/3867082144' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T10:15:46.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:45 vm00 bash[20709]: cluster 2026-03-10T10:15:44.701049+0000 mon.a (mon.0) 188 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:15:46.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:45 vm00 bash[20709]: cluster 2026-03-10T10:15:44.701049+0000 mon.a (mon.0) 188 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:15:46.218 DEBUG:teuthology.orchestra.run.vm03:mon.b> sudo journalctl -f -n 0 -u ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mon.b.service 2026-03-10T10:15:46.219 DEBUG:teuthology.orchestra.run.vm09:mon.c> sudo journalctl -f -n 0 -u ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mon.c.service 2026-03-10T10:15:46.220 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-10T10:15:46.220 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph mon dump -f json 2026-03-10T10:15:47.383 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.c/config 2026-03-10T10:15:47.412 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: cluster 2026-03-10T10:15:45.765126+0000 mgr.a (mgr.14150) 44 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: cluster 2026-03-10T10:15:45.765126+0000 mgr.a (mgr.14150) 44 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: audit 2026-03-10T10:15:46.117481+0000 mgr.a (mgr.14150) 45 : audit [DBG] from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm00:192.168.123.100=a;vm03:192.168.123.103=b;vm09:192.168.123.109=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: audit 2026-03-10T10:15:46.117481+0000 mgr.a (mgr.14150) 45 : audit [DBG] from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm00:192.168.123.100=a;vm03:192.168.123.103=b;vm09:192.168.123.109=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: cephadm 2026-03-10T10:15:46.118555+0000 mgr.a (mgr.14150) 46 : cephadm [INF] Saving service mon spec with placement vm00:192.168.123.100=a;vm03:192.168.123.103=b;vm09:192.168.123.109=c;count:3 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: cephadm 2026-03-10T10:15:46.118555+0000 mgr.a (mgr.14150) 46 : cephadm [INF] Saving service mon spec with placement vm00:192.168.123.100=a;vm03:192.168.123.103=b;vm09:192.168.123.109=c;count:3 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: audit 2026-03-10T10:15:46.121243+0000 mon.a (mon.0) 189 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: audit 2026-03-10T10:15:46.121243+0000 mon.a (mon.0) 189 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: audit 2026-03-10T10:15:46.122082+0000 mon.a (mon.0) 190 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: audit 2026-03-10T10:15:46.122082+0000 mon.a (mon.0) 190 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: audit 2026-03-10T10:15:46.123648+0000 mon.a (mon.0) 191 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: audit 2026-03-10T10:15:46.123648+0000 mon.a (mon.0) 191 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: audit 2026-03-10T10:15:46.124213+0000 mon.a (mon.0) 192 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: audit 2026-03-10T10:15:46.124213+0000 mon.a (mon.0) 192 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: audit 2026-03-10T10:15:46.126949+0000 mon.a (mon.0) 193 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: audit 2026-03-10T10:15:46.126949+0000 mon.a (mon.0) 193 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: audit 2026-03-10T10:15:46.128469+0000 mon.a (mon.0) 194 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: audit 2026-03-10T10:15:46.128469+0000 mon.a (mon.0) 194 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: audit 2026-03-10T10:15:46.128980+0000 mon.a (mon.0) 195 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:47.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: audit 2026-03-10T10:15:46.128980+0000 mon.a (mon.0) 195 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:47.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: cephadm 2026-03-10T10:15:46.129596+0000 mgr.a (mgr.14150) 47 : cephadm [INF] Deploying daemon mon.c on vm09 2026-03-10T10:15:47.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:47 vm00 bash[20709]: cephadm 2026-03-10T10:15:46.129596+0000 mgr.a (mgr.14150) 47 : cephadm [INF] Deploying daemon mon.c on vm09 2026-03-10T10:15:47.686 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:15:47.686 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:15:47.686 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 systemd[1]: Started Ceph mon.c for 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad. 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 0 ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 7 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 0 pidfile_write: ignore empty --pid-file 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 0 load: jerasure load: lrc 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: RocksDB version: 7.9.2 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Git sha 0 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: DB SUMMARY 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: DB Session ID: VR4XBU10FZ19RR463G74 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: CURRENT file: CURRENT 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: MANIFEST file: MANIFEST-000005 size: 59 Bytes 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-c/store.db dir, Total Num: 0, files: 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-c/store.db: 000004.log size: 511 ; 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.error_if_exists: 0 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.create_if_missing: 0 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.paranoid_checks: 1 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.env: 0x55e0e259edc0 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.info_log: 0x55e0e7c30700 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.statistics: (nil) 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.use_fsync: 0 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_log_file_size: 0 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.allow_fallocate: 1 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.use_direct_reads: 0 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-10T10:15:48.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.db_log_dir: 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.wal_dir: 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.write_buffer_manager: 0x55e0e7c35900 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.unordered_write: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.row_cache: None 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.wal_filter: None 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.two_write_queues: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.wal_compression: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.atomic_flush: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.log_readahead_size: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_background_jobs: 2 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_background_compactions: -1 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_subcompactions: 1 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_open_files: -1 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-10T10:15:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_background_flushes: -1 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Compression algorithms supported: 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: kZSTD supported: 0 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: kXpressCompression supported: 0 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: kBZip2Compression supported: 0 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: kLZ4Compression supported: 1 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: kZlibCompression supported: 1 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: kSnappyCompression supported: 1 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-c/store.db/MANIFEST-000005 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.merge_operator: 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compaction_filter: None 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55e0e7c30640) 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cache_index_and_filter_blocks: 1 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: pin_top_level_index_and_filter: 1 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: index_type: 0 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: data_block_index_type: 0 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: index_shortening: 1 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: data_block_hash_table_util_ratio: 0.750000 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: checksum: 4 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: no_block_cache: 0 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: block_cache: 0x55e0e7c57350 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: block_cache_name: BinnedLRUCache 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: block_cache_options: 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: capacity : 536870912 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: num_shard_bits : 4 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: strict_capacity_limit : 0 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: high_pri_pool_ratio: 0.000 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: block_cache_compressed: (nil) 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: persistent_cache: (nil) 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: block_size: 4096 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: block_size_deviation: 10 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: block_restart_interval: 16 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: index_block_restart_interval: 1 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: metadata_block_size: 4096 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: partition_filters: 0 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: use_delta_encoding: 1 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: filter_policy: bloomfilter 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: whole_key_filtering: 1 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: verify_compression: 0 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: read_amp_bytes_per_bit: 0 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: format_version: 5 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: enable_index_compression: 1 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: block_align: 0 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: max_auto_readahead_size: 262144 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: prepopulate_block_cache: 0 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: initial_auto_readahead_size: 8192 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: num_file_reads_for_auto_readahead: 2 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compression: NoCompression 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-10T10:15:48.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.num_levels: 7 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.inplace_update_support: 0 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-10T10:15:48.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.bloom_locality: 0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.max_successive_merges: 0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.ttl: 2592000 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.enable_blob_files: false 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.min_blob_size: 0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.720+0000 7f23f10cbd80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.724+0000 7f23f10cbd80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-c/store.db/MANIFEST-000005 succeeded,manifest_file_number is 5, next_file_number is 7, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.724+0000 7f23f10cbd80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.724+0000 7f23f10cbd80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: dfa3d4be-8016-4700-96eb-1192fbee2322 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.724+0000 7f23f10cbd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773137747726922, "job": 1, "event": "recovery_started", "wal_files": [4]} 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.724+0000 7f23f10cbd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #4 mode 2 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.724+0000 7f23f10cbd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773137747728455, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 8, "file_size": 1643, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 1, "largest_seqno": 5, "table_properties": {"data_size": 523, "index_size": 31, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 115, "raw_average_key_size": 23, "raw_value_size": 401, "raw_average_value_size": 80, "num_data_blocks": 1, "num_entries": 5, "num_filter_entries": 5, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773137747, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "dfa3d4be-8016-4700-96eb-1192fbee2322", "db_session_id": "VR4XBU10FZ19RR463G74", "orig_file_number": 8, "seqno_to_time_mapping": "N/A"}} 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.724+0000 7f23f10cbd80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773137747729132, "job": 1, "event": "recovery_finished"} 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.724+0000 7f23f10cbd80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 10 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.728+0000 7f23f10cbd80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-c/store.db/000004.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.728+0000 7f23f10cbd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55e0e7c58e00 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.728+0000 7f23f10cbd80 4 rocksdb: DB pointer 0x55e0e7d6e000 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.728+0000 7f23e6e95640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.728+0000 7f23e6e95640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: ** DB Stats ** 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: ** Compaction Stats [default] ** 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: L0 1/0 1.60 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.9 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Sum 1/0 1.60 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.9 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.9 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: ** Compaction Stats [default] ** 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.9 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: AddFile(Total Files): cumulative 0, interval 0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: AddFile(Keys): cumulative 0, interval 0 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Cumulative compaction: 0.00 GB write, 0.18 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Interval compaction: 0.00 GB write, 0.18 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-10T10:15:48.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Block cache BinnedLRUCache@0x55e0e7c57350#7 capacity: 512.00 MB usage: 0.22 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 3.2e-05 secs_since: 0 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: Block cache entry stats(count,size,portion): FilterBlock(1,0.11 KB,2.08616e-05%) IndexBlock(1,0.11 KB,2.08616e-05%) Misc(1,0.00 KB,0%) 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: ** File Read Latency Histogram By Level [default] ** 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.732+0000 7f23f10cbd80 0 mon.c does not exist in monmap, will attempt to join an existing cluster 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.732+0000 7f23f10cbd80 0 using public_addr v2:192.168.123.109:0/0 -> [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.732+0000 7f23f10cbd80 0 starting mon.c rank -1 at public addrs [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] at bind addrs [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon_data /var/lib/ceph/mon/ceph-c fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.732+0000 7f23f10cbd80 1 mon.c@-1(???) e0 preinit fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.760+0000 7f23e9e9b640 0 mon.c@-1(synchronizing).mds e1 new map 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.760+0000 7f23e9e9b640 0 mon.c@-1(synchronizing).mds e1 print_map 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: e1 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: btime 2026-03-10T10:14:45:388655+0000 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: legacy client fscid: -1 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: No filesystems configured 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.760+0000 7f23e9e9b640 1 mon.c@-1(synchronizing).osd e0 _set_cache_ratios kv ratio 0.25 inc ratio 0.375 full ratio 0.375 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.760+0000 7f23e9e9b640 1 mon.c@-1(synchronizing).osd e0 register_cache_with_pcm pcm target: 2147483648 pcm max: 1020054732 pcm min: 134217728 inc_osd_cache size: 1 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.760+0000 7f23e9e9b640 1 mon.c@-1(synchronizing).osd e1 e1: 0 total, 0 up, 0 in 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.760+0000 7f23e9e9b640 1 mon.c@-1(synchronizing).osd e2 e2: 0 total, 0 up, 0 in 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.760+0000 7f23e9e9b640 1 mon.c@-1(synchronizing).osd e3 e3: 0 total, 0 up, 0 in 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.760+0000 7f23e9e9b640 1 mon.c@-1(synchronizing).osd e4 e4: 0 total, 0 up, 0 in 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:45.389099+0000 mon.a (mon.0) 0 : cluster [INF] mkfs 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:45.389099+0000 mon.a (mon.0) 0 : cluster [INF] mkfs 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:45.384657+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:45.384657+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.219857+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.219857+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.219882+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.219882+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.219886+0000 mon.a (mon.0) 3 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.219886+0000 mon.a (mon.0) 3 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.219889+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.219889+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.219895+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.219895+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.219898+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.219898+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.219902+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.219902+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.219905+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.219905+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.220108+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.220108+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.220117+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.220117+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.220561+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:46.220561+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:46.258572+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.100:0/624538143' entity='client.admin' 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:46.258572+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.100:0/624538143' entity='client.admin' 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:46.778814+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.100:0/2079901906' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:46.778814+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.100:0/2079901906' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:49.008480+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.100:0/3321573109' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:49.008480+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.100:0/3321573109' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:49.964922+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon a 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:49.964922+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon a 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:49.971198+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: a(active, starting, since 0.00642738s) 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:49.971198+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: a(active, starting, since 0.00642738s) 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:49.973767+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:49.973767+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:49.974443+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:49.974443+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:15:48.012 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:49.974506+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:49.974506+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:49.974553+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:49.974553+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:49.974598+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:49.974598+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:49.983830+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon a is now available 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:49.983830+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon a is now available 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:49.997307+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:49.997307+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:49.999174+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:49.999174+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:50.002748+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:50.002748+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:50.005327+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:50.005327+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:50.006065+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:50.006065+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:50.977179+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: a(active, since 1.0124s) 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:50.977179+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: a(active, since 1.0124s) 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:51.382436+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.100:0/540976360' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:51.382436+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.100:0/540976360' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:51.619140+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.100:0/3131372554' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:51.619140+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.100:0/3131372554' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:51.622788+0000 mon.a (mon.0) 31 : audit [INF] from='client.? 192.168.123.100:0/3131372554' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:51.622788+0000 mon.a (mon.0) 31 : audit [INF] from='client.? 192.168.123.100:0/3131372554' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:51.880558+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.100:0/1053282729' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:51.880558+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.100:0/1053282729' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:52.624524+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.100:0/1053282729' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:52.624524+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.100:0/1053282729' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:52.626178+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:52.626178+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:52.937791+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.100:0/3841211786' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:52.937791+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.100:0/3841211786' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:55.665147+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon a restarted 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:55.665147+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon a restarted 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:55.665541+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon a 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:55.665541+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon a 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:55.669966+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-10T10:15:48.013 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:55.669966+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:55.670031+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e5: a(active, starting, since 0.00457834s) 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:55.670031+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e5: a(active, starting, since 0.00457834s) 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.674294+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.674294+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.674913+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.674913+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.675645+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.675645+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.675986+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.675986+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.676303+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.676303+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:55.681445+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon a is now available 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:55.681445+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon a is now available 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.689780+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.689780+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.692910+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.692910+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.695465+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.695465+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.711238+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.711238+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.713169+0000 mon.a (mon.0) 50 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.713169+0000 mon.a (mon.0) 50 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-debug 2026-03-10T10:15:47.764+0000 7f23e9e9b640 0 mon.c@-1(synchronizing).osd e4 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.764+0000 7f23e9e9b640 0 mon.c@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.764+0000 7f23e9e9b640 0 mon.c@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.764+0000 7f23e9e9b640 0 mon.c@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: 03-10T10:14:55.687288+0000 mgr.a (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:14:55.687288+0000 mgr.a (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.717231+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.717231+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.730672+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:15:48.014 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:55.730672+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:56.084022+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:56.084022+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:56.086584+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:56.086584+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:14:56.600038+0000 mgr.a (mgr.14118) 2 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Bus STARTING 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:14:56.600038+0000 mgr.a (mgr.14118) 2 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Bus STARTING 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:56.673593+0000 mon.a (mon.0) 55 : cluster [DBG] mgrmap e6: a(active, since 1.00814s) 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:56.673593+0000 mon.a (mon.0) 55 : cluster [DBG] mgrmap e6: a(active, since 1.00814s) 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:56.675529+0000 mgr.a (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:56.675529+0000 mgr.a (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:56.679528+0000 mgr.a (mgr.14118) 4 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:56.679528+0000 mgr.a (mgr.14118) 4 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:14:56.707900+0000 mgr.a (mgr.14118) 5 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Serving on https://192.168.123.100:7150 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:14:56.707900+0000 mgr.a (mgr.14118) 5 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Serving on https://192.168.123.100:7150 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:14:56.708300+0000 mgr.a (mgr.14118) 6 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Client ('192.168.123.100', 49166) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:14:56.708300+0000 mgr.a (mgr.14118) 6 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Client ('192.168.123.100', 49166) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:14:56.808912+0000 mgr.a (mgr.14118) 7 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Serving on http://192.168.123.100:8765 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:14:56.808912+0000 mgr.a (mgr.14118) 7 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Serving on http://192.168.123.100:8765 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:14:56.809134+0000 mgr.a (mgr.14118) 8 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Bus STARTED 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:14:56.809134+0000 mgr.a (mgr.14118) 8 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Bus STARTED 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:56.809685+0000 mon.a (mon.0) 56 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:56.809685+0000 mon.a (mon.0) 56 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:56.988109+0000 mgr.a (mgr.14118) 9 : audit [DBG] from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:56.988109+0000 mgr.a (mgr.14118) 9 : audit [DBG] from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:56.992162+0000 mon.a (mon.0) 57 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:56.992162+0000 mon.a (mon.0) 57 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:56.999054+0000 mon.a (mon.0) 58 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:56.999054+0000 mon.a (mon.0) 58 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:57.253328+0000 mgr.a (mgr.14118) 10 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:57.253328+0000 mgr.a (mgr.14118) 10 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:57.525895+0000 mgr.a (mgr.14118) 11 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:57.525895+0000 mgr.a (mgr.14118) 11 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:14:57.526082+0000 mgr.a (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:14:57.526082+0000 mgr.a (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:57.541819+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:57.541819+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:57.543843+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:57.543843+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:57.778687+0000 mgr.a (mgr.14118) 13 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:57.778687+0000 mgr.a (mgr.14118) 13 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:58.023626+0000 mgr.a (mgr.14118) 14 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm00", "addr": "192.168.123.100", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:58.023626+0000 mgr.a (mgr.14118) 14 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm00", "addr": "192.168.123.100", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:58.548356+0000 mon.a (mon.0) 61 : cluster [DBG] mgrmap e7: a(active, since 2s) 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:14:58.548356+0000 mon.a (mon.0) 61 : cluster [DBG] mgrmap e7: a(active, since 2s) 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:14:58.683973+0000 mgr.a (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm00 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:14:58.683973+0000 mgr.a (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm00 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:59.952848+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:59.952848+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:59.953348+0000 mon.a (mon.0) 63 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:14:59.953348+0000 mon.a (mon.0) 63 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:00.289441+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:00.289441+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:14:59.953179+0000 mgr.a (mgr.14118) 16 : cephadm [INF] Added host vm00 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:14:59.953179+0000 mgr.a (mgr.14118) 16 : cephadm [INF] Added host vm00 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:00.285780+0000 mgr.a (mgr.14118) 17 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:00.285780+0000 mgr.a (mgr.14118) 17 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:00.286656+0000 mgr.a (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:00.286656+0000 mgr.a (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:00.772498+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:00.772498+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.015 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:01.102548+0000 mon.a (mon.0) 66 : audit [INF] from='client.? 192.168.123.100:0/1210580903' entity='client.admin' 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:01.102548+0000 mon.a (mon.0) 66 : audit [INF] from='client.? 192.168.123.100:0/1210580903' entity='client.admin' 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:01.414019+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.100:0/3887716780' entity='client.admin' 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:01.414019+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.100:0/3887716780' entity='client.admin' 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:01.467433+0000 mon.a (mon.0) 68 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:01.467433+0000 mon.a (mon.0) 68 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:00.769197+0000 mgr.a (mgr.14118) 19 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:00.769197+0000 mgr.a (mgr.14118) 19 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:00.769897+0000 mgr.a (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:00.769897+0000 mgr.a (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:01.759591+0000 mon.a (mon.0) 69 : audit [INF] from='client.? 192.168.123.100:0/2840361483' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:01.759591+0000 mon.a (mon.0) 69 : audit [INF] from='client.? 192.168.123.100:0/2840361483' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:01.782198+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:01.782198+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:02.635760+0000 mon.a (mon.0) 71 : audit [INF] from='client.? 192.168.123.100:0/2840361483' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:02.635760+0000 mon.a (mon.0) 71 : audit [INF] from='client.? 192.168.123.100:0/2840361483' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:02.638223+0000 mon.a (mon.0) 72 : cluster [DBG] mgrmap e8: a(active, since 6s) 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:02.638223+0000 mon.a (mon.0) 72 : cluster [DBG] mgrmap e8: a(active, since 6s) 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:03.370528+0000 mon.a (mon.0) 73 : audit [DBG] from='client.? 192.168.123.100:0/1190503004' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:03.370528+0000 mon.a (mon.0) 73 : audit [DBG] from='client.? 192.168.123.100:0/1190503004' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:05.753510+0000 mon.a (mon.0) 74 : cluster [INF] Active manager daemon a restarted 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:05.753510+0000 mon.a (mon.0) 74 : cluster [INF] Active manager daemon a restarted 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:05.753731+0000 mon.a (mon.0) 75 : cluster [INF] Activating manager daemon a 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:05.753731+0000 mon.a (mon.0) 75 : cluster [INF] Activating manager daemon a 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:05.759504+0000 mon.a (mon.0) 76 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:05.759504+0000 mon.a (mon.0) 76 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:05.759678+0000 mon.a (mon.0) 77 : cluster [DBG] mgrmap e9: a(active, starting, since 0.00603039s) 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:05.759678+0000 mon.a (mon.0) 77 : cluster [DBG] mgrmap e9: a(active, starting, since 0.00603039s) 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.761353+0000 mon.a (mon.0) 78 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.761353+0000 mon.a (mon.0) 78 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.761760+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.761760+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.762588+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.762588+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.762745+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.762745+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.762934+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.762934+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:05.768935+0000 mon.a (mon.0) 83 : cluster [INF] Manager daemon a is now available 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:05.768935+0000 mon.a (mon.0) 83 : cluster [INF] Manager daemon a is now available 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.781428+0000 mon.a (mon.0) 84 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.781428+0000 mon.a (mon.0) 84 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.791322+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.791322+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.796258+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.796258+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.803146+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:05.803146+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:06.762888+0000 mon.a (mon.0) 88 : cluster [DBG] mgrmap e10: a(active, since 1.00924s) 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:06.762888+0000 mon.a (mon.0) 88 : cluster [DBG] mgrmap e10: a(active, since 1.00924s) 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:06.855107+0000 mgr.a (mgr.14150) 3 : cephadm [INF] [10/Mar/2026:10:15:06] ENGINE Bus STARTING 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:06.855107+0000 mgr.a (mgr.14150) 3 : cephadm [INF] [10/Mar/2026:10:15:06] ENGINE Bus STARTING 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:06.956306+0000 mgr.a (mgr.14150) 4 : cephadm [INF] [10/Mar/2026:10:15:06] ENGINE Serving on http://192.168.123.100:8765 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:06.956306+0000 mgr.a (mgr.14150) 4 : cephadm [INF] [10/Mar/2026:10:15:06] ENGINE Serving on http://192.168.123.100:8765 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:07.026151+0000 mgr.a (mgr.14150) 5 : audit [DBG] from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:07.026151+0000 mgr.a (mgr.14150) 5 : audit [DBG] from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:07.054291+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:07.054291+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:07.056813+0000 mon.a (mon.0) 90 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:07.056813+0000 mon.a (mon.0) 90 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:07.065557+0000 mgr.a (mgr.14150) 6 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Serving on https://192.168.123.100:7150 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:07.065557+0000 mgr.a (mgr.14150) 6 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Serving on https://192.168.123.100:7150 2026-03-10T10:15:48.016 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:07.065605+0000 mgr.a (mgr.14150) 7 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Bus STARTED 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:07.065605+0000 mgr.a (mgr.14150) 7 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Bus STARTED 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:07.065900+0000 mgr.a (mgr.14150) 8 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Client ('192.168.123.100', 40024) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:07.065900+0000 mgr.a (mgr.14150) 8 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Client ('192.168.123.100', 40024) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:07.297971+0000 mgr.a (mgr.14150) 9 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:07.297971+0000 mgr.a (mgr.14150) 9 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:07.447019+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:07.447019+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:07.721638+0000 mon.a (mon.0) 92 : audit [DBG] from='client.? 192.168.123.100:0/1740426705' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:07.721638+0000 mon.a (mon.0) 92 : audit [DBG] from='client.? 192.168.123.100:0/1740426705' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:08.048363+0000 mon.a (mon.0) 93 : audit [INF] from='client.? 192.168.123.100:0/2918787650' entity='client.admin' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:08.048363+0000 mon.a (mon.0) 93 : audit [INF] from='client.? 192.168.123.100:0/2918787650' entity='client.admin' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:08.451766+0000 mon.a (mon.0) 94 : cluster [DBG] mgrmap e11: a(active, since 2s) 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:08.451766+0000 mon.a (mon.0) 94 : cluster [DBG] mgrmap e11: a(active, since 2s) 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:10.693049+0000 mon.a (mon.0) 95 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:10.693049+0000 mon.a (mon.0) 95 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:10.693786+0000 mon.a (mon.0) 96 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:10.693786+0000 mon.a (mon.0) 96 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:10.696540+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:10.696540+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:10.697444+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]: dispatch 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:10.697444+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]: dispatch 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:10.698653+0000 mon.a (mon.0) 99 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]': finished 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:10.698653+0000 mon.a (mon.0) 99 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]': finished 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:11.443247+0000 mgr.a (mgr.14150) 10 : cephadm [INF] Deploying daemon agent.vm00 on vm00 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:11.443247+0000 mgr.a (mgr.14150) 10 : cephadm [INF] Deploying daemon agent.vm00 on vm00 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.072226+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.072226+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.074786+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.074786+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.076813+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.076813+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.082261+0000 mon.a (mon.0) 103 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.082261+0000 mon.a (mon.0) 103 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.085755+0000 mon.a (mon.0) 104 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.085755+0000 mon.a (mon.0) 104 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.092222+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.092222+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.094717+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.094717+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.096816+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.096816+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.099935+0000 mon.a (mon.0) 108 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.099935+0000 mon.a (mon.0) 108 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.106233+0000 mon.a (mon.0) 109 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.106233+0000 mon.a (mon.0) 109 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.109685+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.109685+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.349600+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.349600+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:12.702263+0000 mon.a (mon.0) 112 : cluster [DBG] mgrmap e12: a(active, since 6s) 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:12.702263+0000 mon.a (mon.0) 112 : cluster [DBG] mgrmap e12: a(active, since 6s) 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.797145+0000 mon.a (mon.0) 113 : audit [INF] from='client.? 192.168.123.100:0/3135075601' entity='client.admin' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.797145+0000 mon.a (mon.0) 113 : audit [INF] from='client.? 192.168.123.100:0/3135075601' entity='client.admin' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.802223+0000 mon.a (mon.0) 114 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.802223+0000 mon.a (mon.0) 114 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.807545+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:12.807545+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:15.793178+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:15.793178+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.851569+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.851569+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.017 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.853957+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.853957+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.858528+0000 mgr.a (mgr.14150) 11 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.858528+0000 mgr.a (mgr.14150) 11 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.860037+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.860037+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.860448+0000 mon.a (mon.0) 120 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.860448+0000 mon.a (mon.0) 120 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.861276+0000 mon.a (mon.0) 121 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.861276+0000 mon.a (mon.0) 121 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.861627+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.861627+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:16.862073+0000 mgr.a (mgr.14150) 12 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:16.862073+0000 mgr.a (mgr.14150) 12 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:16.892600+0000 mgr.a (mgr.14150) 13 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:16.892600+0000 mgr.a (mgr.14150) 13 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:16.932604+0000 mgr.a (mgr.14150) 14 : cephadm [INF] Updating vm00:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:16.932604+0000 mgr.a (mgr.14150) 14 : cephadm [INF] Updating vm00:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:16.961644+0000 mgr.a (mgr.14150) 15 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:16.961644+0000 mgr.a (mgr.14150) 15 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.996487+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.996487+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.999145+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:16.999145+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:21.883952+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:21.883952+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:21.886590+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:21.886590+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:22.565126+0000 mgr.a (mgr.14150) 16 : audit [DBG] from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm03", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:22.565126+0000 mgr.a (mgr.14150) 16 : audit [DBG] from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm03", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:23.090899+0000 mgr.a (mgr.14150) 17 : cephadm [INF] Deploying cephadm binary to vm03 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:23.090899+0000 mgr.a (mgr.14150) 17 : cephadm [INF] Deploying cephadm binary to vm03 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:24.284271+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:24.284271+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:24.284686+0000 mgr.a (mgr.14150) 18 : cephadm [INF] Added host vm03 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:24.284686+0000 mgr.a (mgr.14150) 18 : cephadm [INF] Added host vm03 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:24.285038+0000 mon.a (mon.0) 128 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:24.285038+0000 mon.a (mon.0) 128 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:24.557335+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:24.557335+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:25.763446+0000 mgr.a (mgr.14150) 19 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:25.763446+0000 mgr.a (mgr.14150) 19 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.815094+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.815094+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.817730+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.817730+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.818256+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.818256+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.818902+0000 mon.a (mon.0) 133 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.818902+0000 mon.a (mon.0) 133 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.819301+0000 mon.a (mon.0) 134 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.819301+0000 mon.a (mon.0) 134 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:25.819855+0000 mgr.a (mgr.14150) 20 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:25.819855+0000 mgr.a (mgr.14150) 20 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:25.848609+0000 mgr.a (mgr.14150) 21 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:25.848609+0000 mgr.a (mgr.14150) 21 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:25.877234+0000 mgr.a (mgr.14150) 22 : cephadm [INF] Updating vm03:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:25.877234+0000 mgr.a (mgr.14150) 22 : cephadm [INF] Updating vm03:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:25.911845+0000 mgr.a (mgr.14150) 23 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:25.911845+0000 mgr.a (mgr.14150) 23 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.944959+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.944959+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.018 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.947071+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.947071+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.947867+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.947867+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.948841+0000 mon.a (mon.0) 138 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]': finished 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:25.948841+0000 mon.a (mon.0) 138 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]': finished 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:26.326527+0000 mgr.a (mgr.14150) 24 : cephadm [INF] Deploying daemon agent.vm03 on vm03 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:26.326527+0000 mgr.a (mgr.14150) 24 : cephadm [INF] Deploying daemon agent.vm03 on vm03 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:26.971739+0000 mon.a (mon.0) 139 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:26.971739+0000 mon.a (mon.0) 139 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:26.973943+0000 mon.a (mon.0) 140 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:26.973943+0000 mon.a (mon.0) 140 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:26.976613+0000 mon.a (mon.0) 141 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:26.976613+0000 mon.a (mon.0) 141 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:26.986160+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:26.986160+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:26.987509+0000 mon.a (mon.0) 143 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:26.987509+0000 mon.a (mon.0) 143 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:26.988036+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:26.988036+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:26.991688+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:26.991688+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:27.003194+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:27.003194+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:27.004316+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:27.004316+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:27.004722+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:27.004722+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:27.007998+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:27.007998+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:27.231900+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:27.231900+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:27.763612+0000 mgr.a (mgr.14150) 25 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:27.763612+0000 mgr.a (mgr.14150) 25 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:28.276683+0000 mon.a (mon.0) 151 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:28.276683+0000 mon.a (mon.0) 151 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:28.279096+0000 mon.a (mon.0) 152 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:28.279096+0000 mon.a (mon.0) 152 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:29.206326+0000 mgr.a (mgr.14150) 26 : audit [DBG] from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:29.206326+0000 mgr.a (mgr.14150) 26 : audit [DBG] from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:29.763764+0000 mgr.a (mgr.14150) 27 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:29.763764+0000 mgr.a (mgr.14150) 27 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:30.796868+0000 mon.a (mon.0) 153 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:30.796868+0000 mon.a (mon.0) 153 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:31.763932+0000 mgr.a (mgr.14150) 28 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:31.763932+0000 mgr.a (mgr.14150) 28 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:33.256409+0000 mgr.a (mgr.14150) 29 : audit [DBG] from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:33.256409+0000 mgr.a (mgr.14150) 29 : audit [DBG] from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:33.762886+0000 mgr.a (mgr.14150) 30 : cephadm [INF] Deploying cephadm binary to vm09 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:33.762886+0000 mgr.a (mgr.14150) 30 : cephadm [INF] Deploying cephadm binary to vm09 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:33.764093+0000 mgr.a (mgr.14150) 31 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:33.764093+0000 mgr.a (mgr.14150) 31 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:34.955366+0000 mon.a (mon.0) 154 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:34.955366+0000 mon.a (mon.0) 154 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:34.956369+0000 mon.a (mon.0) 155 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:34.956369+0000 mon.a (mon.0) 155 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:35.073114+0000 mon.a (mon.0) 156 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:35.073114+0000 mon.a (mon.0) 156 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:35.076842+0000 mon.a (mon.0) 157 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:35.076842+0000 mon.a (mon.0) 157 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:35.090473+0000 mon.a (mon.0) 158 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:35.090473+0000 mon.a (mon.0) 158 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.019 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:35.240644+0000 mon.a (mon.0) 159 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:35.240644+0000 mon.a (mon.0) 159 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:34.956152+0000 mgr.a (mgr.14150) 32 : cephadm [INF] Added host vm09 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:34.956152+0000 mgr.a (mgr.14150) 32 : cephadm [INF] Added host vm09 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:35.764276+0000 mgr.a (mgr.14150) 33 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:35.764276+0000 mgr.a (mgr.14150) 33 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.498998+0000 mon.a (mon.0) 160 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.498998+0000 mon.a (mon.0) 160 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.500913+0000 mon.a (mon.0) 161 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.500913+0000 mon.a (mon.0) 161 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.501352+0000 mon.a (mon.0) 162 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.501352+0000 mon.a (mon.0) 162 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.501943+0000 mon.a (mon.0) 163 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.501943+0000 mon.a (mon.0) 163 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.502308+0000 mon.a (mon.0) 164 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.502308+0000 mon.a (mon.0) 164 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:36.502878+0000 mgr.a (mgr.14150) 34 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:36.502878+0000 mgr.a (mgr.14150) 34 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:36.529054+0000 mgr.a (mgr.14150) 35 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:36.529054+0000 mgr.a (mgr.14150) 35 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:36.553286+0000 mgr.a (mgr.14150) 36 : cephadm [INF] Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:36.553286+0000 mgr.a (mgr.14150) 36 : cephadm [INF] Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:36.577465+0000 mgr.a (mgr.14150) 37 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:36.577465+0000 mgr.a (mgr.14150) 37 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.604392+0000 mon.a (mon.0) 165 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.604392+0000 mon.a (mon.0) 165 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.606964+0000 mon.a (mon.0) 166 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.606964+0000 mon.a (mon.0) 166 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.607844+0000 mon.a (mon.0) 167 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.607844+0000 mon.a (mon.0) 167 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.608867+0000 mon.a (mon.0) 168 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []}]': finished 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:36.608867+0000 mon.a (mon.0) 168 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []}]': finished 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:37.024492+0000 mgr.a (mgr.14150) 38 : cephadm [INF] Deploying daemon agent.vm09 on vm09 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:37.024492+0000 mgr.a (mgr.14150) 38 : cephadm [INF] Deploying daemon agent.vm09 on vm09 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.640248+0000 mon.a (mon.0) 169 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.640248+0000 mon.a (mon.0) 169 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.643021+0000 mon.a (mon.0) 170 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.643021+0000 mon.a (mon.0) 170 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.645207+0000 mon.a (mon.0) 171 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.645207+0000 mon.a (mon.0) 171 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.660173+0000 mon.a (mon.0) 172 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.660173+0000 mon.a (mon.0) 172 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.661491+0000 mon.a (mon.0) 173 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.661491+0000 mon.a (mon.0) 173 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.661925+0000 mon.a (mon.0) 174 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.661925+0000 mon.a (mon.0) 174 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.665227+0000 mon.a (mon.0) 175 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.665227+0000 mon.a (mon.0) 175 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.678573+0000 mon.a (mon.0) 176 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.678573+0000 mon.a (mon.0) 176 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.679711+0000 mon.a (mon.0) 177 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.679711+0000 mon.a (mon.0) 177 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.680085+0000 mon.a (mon.0) 178 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.680085+0000 mon.a (mon.0) 178 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.682903+0000 mon.a (mon.0) 179 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.682903+0000 mon.a (mon.0) 179 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.908612+0000 mon.a (mon.0) 180 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:37.908612+0000 mon.a (mon.0) 180 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:38.623828+0000 mon.a (mon.0) 181 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:38.623828+0000 mon.a (mon.0) 181 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:38.626247+0000 mon.a (mon.0) 182 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:38.626247+0000 mon.a (mon.0) 182 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:37.764466+0000 mgr.a (mgr.14150) 39 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:37.764466+0000 mgr.a (mgr.14150) 39 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:38.958192+0000 mon.a (mon.0) 183 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:38.958192+0000 mon.a (mon.0) 183 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:38.960842+0000 mon.a (mon.0) 184 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:38.960842+0000 mon.a (mon.0) 184 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:39.545387+0000 mgr.a (mgr.14150) 40 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:39.545387+0000 mgr.a (mgr.14150) 40 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:39.764623+0000 mgr.a (mgr.14150) 41 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:39.764623+0000 mgr.a (mgr.14150) 41 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:40.800636+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:40.800636+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:41.764784+0000 mgr.a (mgr.14150) 42 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:41.764784+0000 mgr.a (mgr.14150) 42 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:44.455176+0000 mon.a (mon.0) 186 : audit [INF] from='client.? 192.168.123.100:0/3867082144' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:44.455176+0000 mon.a (mon.0) 186 : audit [INF] from='client.? 192.168.123.100:0/3867082144' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:43.764968+0000 mgr.a (mgr.14150) 43 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:43.764968+0000 mgr.a (mgr.14150) 43 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:44.699288+0000 mon.a (mon.0) 187 : audit [INF] from='client.? 192.168.123.100:0/3867082144' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:44.699288+0000 mon.a (mon.0) 187 : audit [INF] from='client.? 192.168.123.100:0/3867082144' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:44.701049+0000 mon.a (mon.0) 188 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:44.701049+0000 mon.a (mon.0) 188 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:45.765126+0000 mgr.a (mgr.14150) 44 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cluster 2026-03-10T10:15:45.765126+0000 mgr.a (mgr.14150) 44 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:46.117481+0000 mgr.a (mgr.14150) 45 : audit [DBG] from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm00:192.168.123.100=a;vm03:192.168.123.103=b;vm09:192.168.123.109=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:46.117481+0000 mgr.a (mgr.14150) 45 : audit [DBG] from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm00:192.168.123.100=a;vm03:192.168.123.103=b;vm09:192.168.123.109=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:46.118555+0000 mgr.a (mgr.14150) 46 : cephadm [INF] Saving service mon spec with placement vm00:192.168.123.100=a;vm03:192.168.123.103=b;vm09:192.168.123.109=c;count:3 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:46.118555+0000 mgr.a (mgr.14150) 46 : cephadm [INF] Saving service mon spec with placement vm00:192.168.123.100=a;vm03:192.168.123.103=b;vm09:192.168.123.109=c;count:3 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:46.121243+0000 mon.a (mon.0) 189 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:46.121243+0000 mon.a (mon.0) 189 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:46.122082+0000 mon.a (mon.0) 190 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:46.122082+0000 mon.a (mon.0) 190 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:46.123648+0000 mon.a (mon.0) 191 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:46.123648+0000 mon.a (mon.0) 191 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:46.124213+0000 mon.a (mon.0) 192 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:46.124213+0000 mon.a (mon.0) 192 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:46.126949+0000 mon.a (mon.0) 193 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:46.126949+0000 mon.a (mon.0) 193 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:46.128469+0000 mon.a (mon.0) 194 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:46.128469+0000 mon.a (mon.0) 194 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:46.128980+0000 mon.a (mon.0) 195 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: audit 2026-03-10T10:15:46.128980+0000 mon.a (mon.0) 195 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:46.129596+0000 mgr.a (mgr.14150) 47 : cephadm [INF] Deploying daemon mon.c on vm09 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: cephadm 2026-03-10T10:15:46.129596+0000 mgr.a (mgr.14150) 47 : cephadm [INF] Deploying daemon mon.c on vm09 2026-03-10T10:15:48.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:47 vm09 bash[20916]: debug 2026-03-10T10:15:47.772+0000 7f23e9e9b640 1 mon.c@-1(synchronizing).paxosservice(auth 1..6) refresh upgraded, format 0 -> 3 2026-03-10T10:15:50.387 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 systemd[1]: Started Ceph mon.b for 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad. 2026-03-10T10:15:50.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.446+0000 7fdc31a30d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.446+0000 7fdc31a30d80 0 ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 7 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.446+0000 7fdc31a30d80 0 pidfile_write: ignore empty --pid-file 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 0 load: jerasure load: lrc 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Git sha 0 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Compile date 2026-02-25 18:11:04 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: DB SUMMARY 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: DB Session ID: TB37U2YDLPGV5SP0G42G 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: CURRENT file: CURRENT 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: MANIFEST file: MANIFEST-000005 size: 59 Bytes 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-b/store.db dir, Total Num: 0, files: 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-b/store.db: 000004.log size: 511 ; 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.error_if_exists: 0 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.create_if_missing: 0 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.env: 0x5589c9e88dc0 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.info_log: 0x5589ec98b880 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.statistics: (nil) 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.use_fsync: 0 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.db_log_dir: 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.wal_dir: 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-10T10:15:50.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.write_buffer_manager: 0x5589ec98f900 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.unordered_write: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.row_cache: None 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.wal_filter: None 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.two_write_queues: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.wal_compression: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.atomic_flush: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_open_files: -1 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Compression algorithms supported: 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: kZSTD supported: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: kXpressCompression supported: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: kZlibCompression supported: 1 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-10T10:15:50.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000005 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.merge_operator: 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compaction_filter: None 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5589ec98a480) 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cache_index_and_filter_blocks: 1 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: pin_top_level_index_and_filter: 1 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: index_type: 0 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: data_block_index_type: 0 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: index_shortening: 1 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: data_block_hash_table_util_ratio: 0.750000 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: checksum: 4 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: no_block_cache: 0 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: block_cache: 0x5589ec9b1350 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: block_cache_name: BinnedLRUCache 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: block_cache_options: 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: capacity : 536870912 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: num_shard_bits : 4 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: strict_capacity_limit : 0 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: high_pri_pool_ratio: 0.000 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: block_cache_compressed: (nil) 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: persistent_cache: (nil) 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: block_size: 4096 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: block_size_deviation: 10 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: block_restart_interval: 16 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: index_block_restart_interval: 1 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: metadata_block_size: 4096 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: partition_filters: 0 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: use_delta_encoding: 1 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: filter_policy: bloomfilter 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: whole_key_filtering: 1 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: verify_compression: 0 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: read_amp_bytes_per_bit: 0 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: format_version: 5 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: enable_index_compression: 1 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: block_align: 0 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: max_auto_readahead_size: 262144 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: prepopulate_block_cache: 0 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: initial_auto_readahead_size: 8192 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: num_file_reads_for_auto_readahead: 2 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compression: NoCompression 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.num_levels: 7 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-10T10:15:50.836 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.bloom_locality: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.ttl: 2592000 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-10T10:15:50.837 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.enable_blob_files: false 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.min_blob_size: 0 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000005 succeeded,manifest_file_number is 5, next_file_number is 7, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 70dc415b-5dc8-4161-aa25-85c2cc5b3c30 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773137750453870, "job": 1, "event": "recovery_started", "wal_files": [4]} 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.450+0000 7fdc31a30d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #4 mode 2 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.454+0000 7fdc31a30d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773137750455616, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 8, "file_size": 1643, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 1, "largest_seqno": 5, "table_properties": {"data_size": 523, "index_size": 31, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 115, "raw_average_key_size": 23, "raw_value_size": 401, "raw_average_value_size": 80, "num_data_blocks": 1, "num_entries": 5, "num_filter_entries": 5, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773137750, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "70dc415b-5dc8-4161-aa25-85c2cc5b3c30", "db_session_id": "TB37U2YDLPGV5SP0G42G", "orig_file_number": 8, "seqno_to_time_mapping": "N/A"}} 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.454+0000 7fdc31a30d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1773137750455671, "job": 1, "event": "recovery_finished"} 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.454+0000 7fdc31a30d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 10 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.458+0000 7fdc31a30d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-b/store.db/000004.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.458+0000 7fdc31a30d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x5589ec9b2e00 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.458+0000 7fdc31a30d80 4 rocksdb: DB pointer 0x5589ecabe000 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.458+0000 7fdc31a30d80 0 mon.b does not exist in monmap, will attempt to join an existing cluster 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.458+0000 7fdc31a30d80 0 using public_addr v2:192.168.123.103:0/0 -> [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.458+0000 7fdc31a30d80 0 starting mon.b rank -1 at public addrs [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] at bind addrs [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon_data /var/lib/ceph/mon/ceph-b fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.458+0000 7fdc31a30d80 1 mon.b@-1(???) e0 preinit fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.462+0000 7fdc277fa640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.462+0000 7fdc277fa640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: ** DB Stats ** 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: ** Compaction Stats [default] ** 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-10T10:15:50.838 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: L0 1/0 1.60 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.9 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Sum 1/0 1.60 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.9 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.9 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: ** Compaction Stats [default] ** 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.9 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: AddFile(Total Files): cumulative 0, interval 0 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: AddFile(Keys): cumulative 0, interval 0 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Cumulative compaction: 0.00 GB write, 0.16 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Interval compaction: 0.00 GB write, 0.16 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Block cache BinnedLRUCache@0x5589ec9b1350#7 capacity: 512.00 MB usage: 0.86 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 6e-06 secs_since: 0 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: Block cache entry stats(count,size,portion): DataBlock(1,0.64 KB,0.00012219%) FilterBlock(1,0.11 KB,2.08616e-05%) IndexBlock(1,0.11 KB,2.08616e-05%) Misc(1,0.00 KB,0%) 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: ** File Read Latency Histogram By Level [default] ** 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.478+0000 7fdc2a800640 0 mon.b@-1(synchronizing).mds e1 new map 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.478+0000 7fdc2a800640 0 mon.b@-1(synchronizing).mds e1 print_map 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: e1 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: btime 2026-03-10T10:14:45:388655+0000 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: legacy client fscid: -1 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: No filesystems configured 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.478+0000 7fdc2a800640 1 mon.b@-1(synchronizing).osd e0 _set_cache_ratios kv ratio 0.25 inc ratio 0.375 full ratio 0.375 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.482+0000 7fdc2a800640 1 mon.b@-1(synchronizing).osd e0 register_cache_with_pcm pcm target: 2147483648 pcm max: 1020054732 pcm min: 134217728 inc_osd_cache size: 1 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.482+0000 7fdc2a800640 1 mon.b@-1(synchronizing).osd e1 e1: 0 total, 0 up, 0 in 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.482+0000 7fdc2a800640 1 mon.b@-1(synchronizing).osd e2 e2: 0 total, 0 up, 0 in 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.482+0000 7fdc2a800640 1 mon.b@-1(synchronizing).osd e3 e3: 0 total, 0 up, 0 in 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.482+0000 7fdc2a800640 1 mon.b@-1(synchronizing).osd e4 e4: 0 total, 0 up, 0 in 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:45.389099+0000 mon.adebug 2026-03-10T10:15:50.482+0000 7fdc2a800640 0 mon.b@-1(synchronizing).osd e4 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.482+0000 7fdc2a800640 0 mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.482+0000 7fdc2a800640 0 mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: (mon.0) debug 2026-03-10T10:15:50.482+0000 7fdc2a800640 0 mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: 0 : cluster [INF] mkfs 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:45.389099+0000 mon.a (mon.0) 0 : cluster [INF] mkfs 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:45.384657+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:45.384657+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.219857+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.219857+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.219882+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.219882+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.219886+0000 mon.a (mon.0) 3 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.219886+0000 mon.a (mon.0) 3 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.219889+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.219889+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.219895+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.219895+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.219898+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.219898+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.219902+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.219902+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.219905+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.219905+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.220108+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.220108+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-10T10:15:50.839 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.220117+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.220117+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.220561+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:46.220561+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:46.258572+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.100:0/624538143' entity='client.admin' 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:46.258572+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.100:0/624538143' entity='client.admin' 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:46.778814+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.100:0/2079901906' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:46.778814+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.100:0/2079901906' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:49.008480+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.100:0/3321573109' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:49.008480+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.100:0/3321573109' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:49.964922+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon a 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:49.964922+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon a 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:49.971198+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: a(active, starting, since 0.00642738s) 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:49.971198+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: a(active, starting, since 0.00642738s) 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:49.973767+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:49.973767+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:49.974443+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:49.974443+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:49.974506+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:49.974506+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:49.974553+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:49.974553+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:49.974598+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:49.974598+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:49.983830+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon a is now available 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:49.983830+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon a is now available 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:49.997307+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:49.997307+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:49.999174+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:49.999174+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:50.002748+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:50.002748+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:50.005327+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:50.005327+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:50.006065+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:50.006065+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.100:0/3289954527' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:50.977179+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: a(active, since 1.0124s) 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:50.977179+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: a(active, since 1.0124s) 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:51.382436+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.100:0/540976360' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:51.382436+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.100:0/540976360' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:51.619140+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.100:0/3131372554' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:51.619140+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.100:0/3131372554' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:51.622788+0000 mon.a (mon.0) 31 : audit [INF] from='client.? 192.168.123.100:0/3131372554' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:51.622788+0000 mon.a (mon.0) 31 : audit [INF] from='client.? 192.168.123.100:0/3131372554' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:51.880558+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.100:0/1053282729' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:51.880558+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.100:0/1053282729' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:52.624524+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.100:0/1053282729' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:52.624524+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.100:0/1053282729' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:52.626178+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:52.626178+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e4: a(active, since 2s) 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:52.937791+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.100:0/3841211786' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:52.937791+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.100:0/3841211786' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:55.665147+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon a restarted 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:55.665147+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon a restarted 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:55.665541+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon a 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:55.665541+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon a 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:55.669966+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-10T10:15:50.840 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:55.669966+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:55.670031+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e5: a(active, starting, since 0.00457834s) 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:55.670031+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e5: a(active, starting, since 0.00457834s) 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.674294+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.674294+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.674913+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.674913+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.675645+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.675645+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.675986+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.675986+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.676303+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.676303+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:55.681445+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon a is now available 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:55.681445+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon a is now available 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.689780+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.689780+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.692910+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.692910+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.695465+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.695465+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.711238+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.711238+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.713169+0000 mon.a (mon.0) 50 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.713169+0000 mon.a (mon.0) 50 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:55.687288+0000 mgr.a (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:55.687288+0000 mgr.a (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.717231+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.717231+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.730672+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:55.730672+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:56.084022+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:56.084022+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:56.086584+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:56.086584+0000 mon.a (mon.0) 54 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:56.600038+0000 mgr.a (mgr.14118) 2 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Bus STARTING 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:56.600038+0000 mgr.a (mgr.14118) 2 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Bus STARTING 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:56.673593+0000 mon.a (mon.0) 55 : cluster [DBG] mgrmap e6: a(active, since 1.00814s) 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:56.673593+0000 mon.a (mon.0) 55 : cluster [DBG] mgrmap e6: a(active, since 1.00814s) 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:56.675529+0000 mgr.a (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:56.675529+0000 mgr.a (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:56.679528+0000 mgr.a (mgr.14118) 4 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:56.679528+0000 mgr.a (mgr.14118) 4 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:56.707900+0000 mgr.a (mgr.14118) 5 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Serving on https://192.168.123.100:7150 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:56.707900+0000 mgr.a (mgr.14118) 5 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Serving on https://192.168.123.100:7150 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:56.708300+0000 mgr.a (mgr.14118) 6 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Client ('192.168.123.100', 49166) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:56.708300+0000 mgr.a (mgr.14118) 6 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Client ('192.168.123.100', 49166) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:56.808912+0000 mgr.a (mgr.14118) 7 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Serving on http://192.168.123.100:8765 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:56.808912+0000 mgr.a (mgr.14118) 7 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Serving on http://192.168.123.100:8765 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:56.809134+0000 mgr.a (mgr.14118) 8 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Bus STARTED 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:56.809134+0000 mgr.a (mgr.14118) 8 : cephadm [INF] [10/Mar/2026:10:14:56] ENGINE Bus STARTED 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:56.809685+0000 mon.a (mon.0) 56 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:56.809685+0000 mon.a (mon.0) 56 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:56.988109+0000 mgr.a (mgr.14118) 9 : audit [DBG] from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:56.988109+0000 mgr.a (mgr.14118) 9 : audit [DBG] from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:56.992162+0000 mon.a (mon.0) 57 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:56.992162+0000 mon.a (mon.0) 57 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:56.999054+0000 mon.a (mon.0) 58 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:56.999054+0000 mon.a (mon.0) 58 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:57.253328+0000 mgr.a (mgr.14118) 10 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:57.253328+0000 mgr.a (mgr.14118) 10 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:57.525895+0000 mgr.a (mgr.14118) 11 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.841 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:57.525895+0000 mgr.a (mgr.14118) 11 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:57.526082+0000 mgr.a (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:57.526082+0000 mgr.a (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:57.541819+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:57.541819+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:57.543843+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:57.543843+0000 mon.a (mon.0) 60 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:57.778687+0000 mgr.a (mgr.14118) 13 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:57.778687+0000 mgr.a (mgr.14118) 13 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:58.023626+0000 mgr.a (mgr.14118) 14 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm00", "addr": "192.168.123.100", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:58.023626+0000 mgr.a (mgr.14118) 14 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm00", "addr": "192.168.123.100", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:58.548356+0000 mon.a (mon.0) 61 : cluster [DBG] mgrmap e7: a(active, since 2s) 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:14:58.548356+0000 mon.a (mon.0) 61 : cluster [DBG] mgrmap e7: a(active, since 2s) 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:58.683973+0000 mgr.a (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm00 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:58.683973+0000 mgr.a (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm00 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:59.952848+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:59.952848+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:59.953348+0000 mon.a (mon.0) 63 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:14:59.953348+0000 mon.a (mon.0) 63 : audit [DBG] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:00.289441+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:00.289441+0000 mon.a (mon.0) 64 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:59.953179+0000 mgr.a (mgr.14118) 16 : cephadm [INF] Added host vm00 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:14:59.953179+0000 mgr.a (mgr.14118) 16 : cephadm [INF] Added host vm00 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:00.285780+0000 mgr.a (mgr.14118) 17 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:00.285780+0000 mgr.a (mgr.14118) 17 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:00.286656+0000 mgr.a (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:00.286656+0000 mgr.a (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:00.772498+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:00.772498+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:01.102548+0000 mon.a (mon.0) 66 : audit [INF] from='client.? 192.168.123.100:0/1210580903' entity='client.admin' 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:01.102548+0000 mon.a (mon.0) 66 : audit [INF] from='client.? 192.168.123.100:0/1210580903' entity='client.admin' 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:01.414019+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.100:0/3887716780' entity='client.admin' 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:01.414019+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.100:0/3887716780' entity='client.admin' 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:01.467433+0000 mon.a (mon.0) 68 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:01.467433+0000 mon.a (mon.0) 68 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:00.769197+0000 mgr.a (mgr.14118) 19 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:00.769197+0000 mgr.a (mgr.14118) 19 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:00.769897+0000 mgr.a (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:00.769897+0000 mgr.a (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:01.759591+0000 mon.a (mon.0) 69 : audit [INF] from='client.? 192.168.123.100:0/2840361483' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T10:15:50.842 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:01.759591+0000 mon.a (mon.0) 69 : audit [INF] from='client.? 192.168.123.100:0/2840361483' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:01.782198+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:01.782198+0000 mon.a (mon.0) 70 : audit [INF] from='mgr.14118 192.168.123.100:0/3296121846' entity='mgr.a' 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:02.635760+0000 mon.a (mon.0) 71 : audit [INF] from='client.? 192.168.123.100:0/2840361483' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:02.635760+0000 mon.a (mon.0) 71 : audit [INF] from='client.? 192.168.123.100:0/2840361483' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:02.638223+0000 mon.a (mon.0) 72 : cluster [DBG] mgrmap e8: a(active, since 6s) 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:02.638223+0000 mon.a (mon.0) 72 : cluster [DBG] mgrmap e8: a(active, since 6s) 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:03.370528+0000 mon.a (mon.0) 73 : audit [DBG] from='client.? 192.168.123.100:0/1190503004' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:03.370528+0000 mon.a (mon.0) 73 : audit [DBG] from='client.? 192.168.123.100:0/1190503004' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:05.753510+0000 mon.a (mon.0) 74 : cluster [INF] Active manager daemon a restarted 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:05.753510+0000 mon.a (mon.0) 74 : cluster [INF] Active manager daemon a restarted 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:05.753731+0000 mon.a (mon.0) 75 : cluster [INF] Activating manager daemon a 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:05.753731+0000 mon.a (mon.0) 75 : cluster [INF] Activating manager daemon a 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:05.759504+0000 mon.a (mon.0) 76 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:05.759504+0000 mon.a (mon.0) 76 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:05.759678+0000 mon.a (mon.0) 77 : cluster [DBG] mgrmap e9: a(active, starting, since 0.00603039s) 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:05.759678+0000 mon.a (mon.0) 77 : cluster [DBG] mgrmap e9: a(active, starting, since 0.00603039s) 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.761353+0000 mon.a (mon.0) 78 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.761353+0000 mon.a (mon.0) 78 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.761760+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.761760+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.762588+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.762588+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.762745+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.762745+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.762934+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.762934+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:05.768935+0000 mon.a (mon.0) 83 : cluster [INF] Manager daemon a is now available 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:05.768935+0000 mon.a (mon.0) 83 : cluster [INF] Manager daemon a is now available 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.781428+0000 mon.a (mon.0) 84 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.781428+0000 mon.a (mon.0) 84 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.791322+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.791322+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.796258+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.796258+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.803146+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:05.803146+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:06.762888+0000 mon.a (mon.0) 88 : cluster [DBG] mgrmap e10: a(active, since 1.00924s) 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:06.762888+0000 mon.a (mon.0) 88 : cluster [DBG] mgrmap e10: a(active, since 1.00924s) 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:06.855107+0000 mgr.a (mgr.14150) 3 : cephadm [INF] [10/Mar/2026:10:15:06] ENGINE Bus STARTING 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:06.855107+0000 mgr.a (mgr.14150) 3 : cephadm [INF] [10/Mar/2026:10:15:06] ENGINE Bus STARTING 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:06.956306+0000 mgr.a (mgr.14150) 4 : cephadm [INF] [10/Mar/2026:10:15:06] ENGINE Serving on http://192.168.123.100:8765 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:06.956306+0000 mgr.a (mgr.14150) 4 : cephadm [INF] [10/Mar/2026:10:15:06] ENGINE Serving on http://192.168.123.100:8765 2026-03-10T10:15:50.843 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:07.026151+0000 mgr.a (mgr.14150) 5 : audit [DBG] from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:07.026151+0000 mgr.a (mgr.14150) 5 : audit [DBG] from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:07.054291+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:07.054291+0000 mon.a (mon.0) 89 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:07.056813+0000 mon.a (mon.0) 90 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:07.056813+0000 mon.a (mon.0) 90 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:07.065557+0000 mgr.a (mgr.14150) 6 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Serving on https://192.168.123.100:7150 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:07.065557+0000 mgr.a (mgr.14150) 6 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Serving on https://192.168.123.100:7150 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:07.065605+0000 mgr.a (mgr.14150) 7 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Bus STARTED 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:07.065605+0000 mgr.a (mgr.14150) 7 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Bus STARTED 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:07.065900+0000 mgr.a (mgr.14150) 8 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Client ('192.168.123.100', 40024) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:07.065900+0000 mgr.a (mgr.14150) 8 : cephadm [INF] [10/Mar/2026:10:15:07] ENGINE Client ('192.168.123.100', 40024) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:07.297971+0000 mgr.a (mgr.14150) 9 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:07.297971+0000 mgr.a (mgr.14150) 9 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:07.447019+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:07.447019+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:07.721638+0000 mon.a (mon.0) 92 : audit [DBG] from='client.? 192.168.123.100:0/1740426705' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:07.721638+0000 mon.a (mon.0) 92 : audit [DBG] from='client.? 192.168.123.100:0/1740426705' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:08.048363+0000 mon.a (mon.0) 93 : audit [INF] from='client.? 192.168.123.100:0/2918787650' entity='client.admin' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:08.048363+0000 mon.a (mon.0) 93 : audit [INF] from='client.? 192.168.123.100:0/2918787650' entity='client.admin' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:08.451766+0000 mon.a (mon.0) 94 : cluster [DBG] mgrmap e11: a(active, since 2s) 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:08.451766+0000 mon.a (mon.0) 94 : cluster [DBG] mgrmap e11: a(active, since 2s) 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:10.693049+0000 mon.a (mon.0) 95 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:10.693049+0000 mon.a (mon.0) 95 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:10.693786+0000 mon.a (mon.0) 96 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:10.693786+0000 mon.a (mon.0) 96 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:10.696540+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:10.696540+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:10.697444+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]: dispatch 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:10.697444+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]: dispatch 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:10.698653+0000 mon.a (mon.0) 99 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]': finished 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:10.698653+0000 mon.a (mon.0) 99 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]': finished 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:11.443247+0000 mgr.a (mgr.14150) 10 : cephadm [INF] Deploying daemon agent.vm00 on vm00 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:11.443247+0000 mgr.a (mgr.14150) 10 : cephadm [INF] Deploying daemon agent.vm00 on vm00 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.072226+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.072226+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.074786+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.074786+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.076813+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.076813+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.082261+0000 mon.a (mon.0) 103 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.082261+0000 mon.a (mon.0) 103 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.085755+0000 mon.a (mon.0) 104 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.085755+0000 mon.a (mon.0) 104 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.092222+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.092222+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.094717+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.094717+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.096816+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.096816+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.099935+0000 mon.a (mon.0) 108 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.099935+0000 mon.a (mon.0) 108 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.106233+0000 mon.a (mon.0) 109 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.106233+0000 mon.a (mon.0) 109 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.109685+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.109685+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.349600+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.349600+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:12.702263+0000 mon.a (mon.0) 112 : cluster [DBG] mgrmap e12: a(active, since 6s) 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:12.702263+0000 mon.a (mon.0) 112 : cluster [DBG] mgrmap e12: a(active, since 6s) 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.797145+0000 mon.a (mon.0) 113 : audit [INF] from='client.? 192.168.123.100:0/3135075601' entity='client.admin' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.797145+0000 mon.a (mon.0) 113 : audit [INF] from='client.? 192.168.123.100:0/3135075601' entity='client.admin' 2026-03-10T10:15:50.844 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.802223+0000 mon.a (mon.0) 114 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.802223+0000 mon.a (mon.0) 114 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.807545+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:12.807545+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:15.793178+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:15.793178+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.851569+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.851569+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.853957+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.853957+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.858528+0000 mgr.a (mgr.14150) 11 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.858528+0000 mgr.a (mgr.14150) 11 : audit [DBG] from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.860037+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.860037+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.860448+0000 mon.a (mon.0) 120 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.860448+0000 mon.a (mon.0) 120 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.861276+0000 mon.a (mon.0) 121 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.861276+0000 mon.a (mon.0) 121 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.861627+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.861627+0000 mon.a (mon.0) 122 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:16.862073+0000 mgr.a (mgr.14150) 12 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:16.862073+0000 mgr.a (mgr.14150) 12 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:16.892600+0000 mgr.a (mgr.14150) 13 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:16.892600+0000 mgr.a (mgr.14150) 13 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:16.932604+0000 mgr.a (mgr.14150) 14 : cephadm [INF] Updating vm00:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:16.932604+0000 mgr.a (mgr.14150) 14 : cephadm [INF] Updating vm00:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:16.961644+0000 mgr.a (mgr.14150) 15 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:16.961644+0000 mgr.a (mgr.14150) 15 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.996487+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.996487+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.999145+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:16.999145+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:21.883952+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:21.883952+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:21.886590+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:21.886590+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:22.565126+0000 mgr.a (mgr.14150) 16 : audit [DBG] from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm03", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:22.565126+0000 mgr.a (mgr.14150) 16 : audit [DBG] from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm03", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:23.090899+0000 mgr.a (mgr.14150) 17 : cephadm [INF] Deploying cephadm binary to vm03 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:23.090899+0000 mgr.a (mgr.14150) 17 : cephadm [INF] Deploying cephadm binary to vm03 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:24.284271+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:24.284271+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:24.284686+0000 mgr.a (mgr.14150) 18 : cephadm [INF] Added host vm03 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:24.284686+0000 mgr.a (mgr.14150) 18 : cephadm [INF] Added host vm03 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:24.285038+0000 mon.a (mon.0) 128 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:24.285038+0000 mon.a (mon.0) 128 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:24.557335+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:24.557335+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:25.763446+0000 mgr.a (mgr.14150) 19 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:25.763446+0000 mgr.a (mgr.14150) 19 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.815094+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.815094+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.817730+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.817730+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.818256+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.818256+0000 mon.a (mon.0) 132 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.818902+0000 mon.a (mon.0) 133 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.818902+0000 mon.a (mon.0) 133 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.819301+0000 mon.a (mon.0) 134 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.819301+0000 mon.a (mon.0) 134 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:25.819855+0000 mgr.a (mgr.14150) 20 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:25.819855+0000 mgr.a (mgr.14150) 20 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:25.848609+0000 mgr.a (mgr.14150) 21 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:25.848609+0000 mgr.a (mgr.14150) 21 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:50.845 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:25.877234+0000 mgr.a (mgr.14150) 22 : cephadm [INF] Updating vm03:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:25.877234+0000 mgr.a (mgr.14150) 22 : cephadm [INF] Updating vm03:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:25.911845+0000 mgr.a (mgr.14150) 23 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:25.911845+0000 mgr.a (mgr.14150) 23 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.944959+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.944959+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.947071+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.947071+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.947867+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.947867+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.948841+0000 mon.a (mon.0) 138 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]': finished 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:25.948841+0000 mon.a (mon.0) 138 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]': finished 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:26.326527+0000 mgr.a (mgr.14150) 24 : cephadm [INF] Deploying daemon agent.vm03 on vm03 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:26.326527+0000 mgr.a (mgr.14150) 24 : cephadm [INF] Deploying daemon agent.vm03 on vm03 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:26.971739+0000 mon.a (mon.0) 139 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:26.971739+0000 mon.a (mon.0) 139 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:26.973943+0000 mon.a (mon.0) 140 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:26.973943+0000 mon.a (mon.0) 140 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:26.976613+0000 mon.a (mon.0) 141 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:26.976613+0000 mon.a (mon.0) 141 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:26.986160+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:26.986160+0000 mon.a (mon.0) 142 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:26.987509+0000 mon.a (mon.0) 143 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:26.987509+0000 mon.a (mon.0) 143 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:26.988036+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:26.988036+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:26.991688+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:26.991688+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:27.003194+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:27.003194+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:27.004316+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:27.004316+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:27.004722+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:27.004722+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:27.007998+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:27.007998+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:27.231900+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:27.231900+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:27.763612+0000 mgr.a (mgr.14150) 25 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:27.763612+0000 mgr.a (mgr.14150) 25 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:28.276683+0000 mon.a (mon.0) 151 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:28.276683+0000 mon.a (mon.0) 151 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:28.279096+0000 mon.a (mon.0) 152 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:28.279096+0000 mon.a (mon.0) 152 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:29.206326+0000 mgr.a (mgr.14150) 26 : audit [DBG] from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:29.206326+0000 mgr.a (mgr.14150) 26 : audit [DBG] from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:29.763764+0000 mgr.a (mgr.14150) 27 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:29.763764+0000 mgr.a (mgr.14150) 27 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:30.796868+0000 mon.a (mon.0) 153 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:30.796868+0000 mon.a (mon.0) 153 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:31.763932+0000 mgr.a (mgr.14150) 28 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:31.763932+0000 mgr.a (mgr.14150) 28 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:33.256409+0000 mgr.a (mgr.14150) 29 : audit [DBG] from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:33.256409+0000 mgr.a (mgr.14150) 29 : audit [DBG] from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:33.762886+0000 mgr.a (mgr.14150) 30 : cephadm [INF] Deploying cephadm binary to vm09 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:33.762886+0000 mgr.a (mgr.14150) 30 : cephadm [INF] Deploying cephadm binary to vm09 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:33.764093+0000 mgr.a (mgr.14150) 31 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:33.764093+0000 mgr.a (mgr.14150) 31 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:34.955366+0000 mon.a (mon.0) 154 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:34.955366+0000 mon.a (mon.0) 154 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:34.956369+0000 mon.a (mon.0) 155 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:34.956369+0000 mon.a (mon.0) 155 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:35.073114+0000 mon.a (mon.0) 156 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.846 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:35.073114+0000 mon.a (mon.0) 156 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:35.076842+0000 mon.a (mon.0) 157 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:35.076842+0000 mon.a (mon.0) 157 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:35.090473+0000 mon.a (mon.0) 158 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:35.090473+0000 mon.a (mon.0) 158 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:35.240644+0000 mon.a (mon.0) 159 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:35.240644+0000 mon.a (mon.0) 159 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:34.956152+0000 mgr.a (mgr.14150) 32 : cephadm [INF] Added host vm09 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:34.956152+0000 mgr.a (mgr.14150) 32 : cephadm [INF] Added host vm09 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:35.764276+0000 mgr.a (mgr.14150) 33 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:35.764276+0000 mgr.a (mgr.14150) 33 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.498998+0000 mon.a (mon.0) 160 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.498998+0000 mon.a (mon.0) 160 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.500913+0000 mon.a (mon.0) 161 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.500913+0000 mon.a (mon.0) 161 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.501352+0000 mon.a (mon.0) 162 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.501352+0000 mon.a (mon.0) 162 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.501943+0000 mon.a (mon.0) 163 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.501943+0000 mon.a (mon.0) 163 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.502308+0000 mon.a (mon.0) 164 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.502308+0000 mon.a (mon.0) 164 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:36.502878+0000 mgr.a (mgr.14150) 34 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:36.502878+0000 mgr.a (mgr.14150) 34 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:36.529054+0000 mgr.a (mgr.14150) 35 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:36.529054+0000 mgr.a (mgr.14150) 35 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:36.553286+0000 mgr.a (mgr.14150) 36 : cephadm [INF] Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:36.553286+0000 mgr.a (mgr.14150) 36 : cephadm [INF] Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:36.577465+0000 mgr.a (mgr.14150) 37 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:36.577465+0000 mgr.a (mgr.14150) 37 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.client.admin.keyring 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.604392+0000 mon.a (mon.0) 165 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.604392+0000 mon.a (mon.0) 165 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.606964+0000 mon.a (mon.0) 166 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.606964+0000 mon.a (mon.0) 166 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.607844+0000 mon.a (mon.0) 167 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.607844+0000 mon.a (mon.0) 167 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.608867+0000 mon.a (mon.0) 168 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []}]': finished 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:36.608867+0000 mon.a (mon.0) 168 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm09", "caps": []}]': finished 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:37.024492+0000 mgr.a (mgr.14150) 38 : cephadm [INF] Deploying daemon agent.vm09 on vm09 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:37.024492+0000 mgr.a (mgr.14150) 38 : cephadm [INF] Deploying daemon agent.vm09 on vm09 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.640248+0000 mon.a (mon.0) 169 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.640248+0000 mon.a (mon.0) 169 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.643021+0000 mon.a (mon.0) 170 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.643021+0000 mon.a (mon.0) 170 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.645207+0000 mon.a (mon.0) 171 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.645207+0000 mon.a (mon.0) 171 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.660173+0000 mon.a (mon.0) 172 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.660173+0000 mon.a (mon.0) 172 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.661491+0000 mon.a (mon.0) 173 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.661491+0000 mon.a (mon.0) 173 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.661925+0000 mon.a (mon.0) 174 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.661925+0000 mon.a (mon.0) 174 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.665227+0000 mon.a (mon.0) 175 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.665227+0000 mon.a (mon.0) 175 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.678573+0000 mon.a (mon.0) 176 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.678573+0000 mon.a (mon.0) 176 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.679711+0000 mon.a (mon.0) 177 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.679711+0000 mon.a (mon.0) 177 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.680085+0000 mon.a (mon.0) 178 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.680085+0000 mon.a (mon.0) 178 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.682903+0000 mon.a (mon.0) 179 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.682903+0000 mon.a (mon.0) 179 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.908612+0000 mon.a (mon.0) 180 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:37.908612+0000 mon.a (mon.0) 180 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.847 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:38.623828+0000 mon.a (mon.0) 181 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:38.623828+0000 mon.a (mon.0) 181 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:38.626247+0000 mon.a (mon.0) 182 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:38.626247+0000 mon.a (mon.0) 182 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:37.764466+0000 mgr.a (mgr.14150) 39 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:37.764466+0000 mgr.a (mgr.14150) 39 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:38.958192+0000 mon.a (mon.0) 183 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:38.958192+0000 mon.a (mon.0) 183 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:38.960842+0000 mon.a (mon.0) 184 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:38.960842+0000 mon.a (mon.0) 184 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:39.545387+0000 mgr.a (mgr.14150) 40 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:39.545387+0000 mgr.a (mgr.14150) 40 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:39.764623+0000 mgr.a (mgr.14150) 41 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:39.764623+0000 mgr.a (mgr.14150) 41 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:40.800636+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:40.800636+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:41.764784+0000 mgr.a (mgr.14150) 42 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:41.764784+0000 mgr.a (mgr.14150) 42 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:44.455176+0000 mon.a (mon.0) 186 : audit [INF] from='client.? 192.168.123.100:0/3867082144' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:44.455176+0000 mon.a (mon.0) 186 : audit [INF] from='client.? 192.168.123.100:0/3867082144' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:43.764968+0000 mgr.a (mgr.14150) 43 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:43.764968+0000 mgr.a (mgr.14150) 43 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:44.699288+0000 mon.a (mon.0) 187 : audit [INF] from='client.? 192.168.123.100:0/3867082144' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:44.699288+0000 mon.a (mon.0) 187 : audit [INF] from='client.? 192.168.123.100:0/3867082144' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:44.701049+0000 mon.a (mon.0) 188 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:44.701049+0000 mon.a (mon.0) 188 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:45.765126+0000 mgr.a (mgr.14150) 44 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cluster 2026-03-10T10:15:45.765126+0000 mgr.a (mgr.14150) 44 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:46.117481+0000 mgr.a (mgr.14150) 45 : audit [DBG] from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm00:192.168.123.100=a;vm03:192.168.123.103=b;vm09:192.168.123.109=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:46.117481+0000 mgr.a (mgr.14150) 45 : audit [DBG] from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm00:192.168.123.100=a;vm03:192.168.123.103=b;vm09:192.168.123.109=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:46.118555+0000 mgr.a (mgr.14150) 46 : cephadm [INF] Saving service mon spec with placement vm00:192.168.123.100=a;vm03:192.168.123.103=b;vm09:192.168.123.109=c;count:3 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:46.118555+0000 mgr.a (mgr.14150) 46 : cephadm [INF] Saving service mon spec with placement vm00:192.168.123.100=a;vm03:192.168.123.103=b;vm09:192.168.123.109=c;count:3 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:46.121243+0000 mon.a (mon.0) 189 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:46.121243+0000 mon.a (mon.0) 189 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:46.122082+0000 mon.a (mon.0) 190 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:46.122082+0000 mon.a (mon.0) 190 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:46.123648+0000 mon.a (mon.0) 191 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:46.123648+0000 mon.a (mon.0) 191 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:46.124213+0000 mon.a (mon.0) 192 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:46.124213+0000 mon.a (mon.0) 192 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:46.126949+0000 mon.a (mon.0) 193 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:46.126949+0000 mon.a (mon.0) 193 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:46.128469+0000 mon.a (mon.0) 194 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:46.128469+0000 mon.a (mon.0) 194 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:46.128980+0000 mon.a (mon.0) 195 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: audit 2026-03-10T10:15:46.128980+0000 mon.a (mon.0) 195 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:46.129596+0000 mgr.a (mgr.14150) 47 : cephadm [INF] Deploying daemon mon.c on vm09 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: cephadm 2026-03-10T10:15:46.129596+0000 mgr.a (mgr.14150) 47 : cephadm [INF] Deploying daemon mon.c on vm09 2026-03-10T10:15:50.848 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:50 vm03 bash[21274]: debug 2026-03-10T10:15:50.518+0000 7fdc2a800640 1 mon.b@-1(synchronizing).paxosservice(auth 1..6) refresh upgraded, format 0 -> 3 2026-03-10T10:15:52.794 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 2 2026-03-10T10:15:52.794 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:15:52.794 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":2,"fsid":"9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad","modified":"2026-03-10T10:15:47.779518Z","created":"2026-03-10T10:14:44.248116Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:3300","nonce":0},{"type":"v1","addr":"192.168.123.100:6789","nonce":0}]},"addr":"192.168.123.100:6789/0","public_addr":"192.168.123.100:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:3300","nonce":0},{"type":"v1","addr":"192.168.123.109:6789","nonce":0}]},"addr":"192.168.123.109:6789/0","public_addr":"192.168.123.109:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:47.765280+0000 mgr.a (mgr.14150) 49 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:47.765280+0000 mgr.a (mgr.14150) 49 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:47.782565+0000 mon.a (mon.0) 201 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:47.782565+0000 mon.a (mon.0) 201 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:47.782680+0000 mon.a (mon.0) 202 : cluster [INF] mon.a calling monitor election 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:47.782680+0000 mon.a (mon.0) 202 : cluster [INF] mon.a calling monitor election 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:47.783736+0000 mon.a (mon.0) 203 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:47.783736+0000 mon.a (mon.0) 203 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:47.908047+0000 mon.a (mon.0) 204 : audit [DBG] from='client.? 192.168.123.109:0/3793675847' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:47.908047+0000 mon.a (mon.0) 204 : audit [DBG] from='client.? 192.168.123.109:0/3793675847' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:48.777060+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:48.777060+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:49.765443+0000 mgr.a (mgr.14150) 50 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:49.765443+0000 mgr.a (mgr.14150) 50 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:49.777615+0000 mon.a (mon.0) 206 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:49.777615+0000 mon.a (mon.0) 206 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:49.784703+0000 mon.c (mon.1) 1 : cluster [INF] mon.c calling monitor election 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:49.784703+0000 mon.c (mon.1) 1 : cluster [INF] mon.c calling monitor election 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:50.521397+0000 mon.a (mon.0) 207 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:50.521397+0000 mon.a (mon.0) 207 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:50.777581+0000 mon.a (mon.0) 208 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:50.777581+0000 mon.a (mon.0) 208 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:51.521436+0000 mon.a (mon.0) 209 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:51.521436+0000 mon.a (mon.0) 209 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:51.765607+0000 mgr.a (mgr.14150) 51 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:51.765607+0000 mgr.a (mgr.14150) 51 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:51.777427+0000 mon.a (mon.0) 210 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:51.777427+0000 mon.a (mon.0) 210 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.522097+0000 mon.a (mon.0) 211 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.522097+0000 mon.a (mon.0) 211 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.777890+0000 mon.a (mon.0) 212 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.777890+0000 mon.a (mon.0) 212 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.787542+0000 mon.a (mon.0) 213 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.787542+0000 mon.a (mon.0) 213 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791023+0000 mon.a (mon.0) 214 : cluster [DBG] monmap epoch 2 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791023+0000 mon.a (mon.0) 214 : cluster [DBG] monmap epoch 2 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791049+0000 mon.a (mon.0) 215 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791049+0000 mon.a (mon.0) 215 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791062+0000 mon.a (mon.0) 216 : cluster [DBG] last_changed 2026-03-10T10:15:47.779518+0000 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791062+0000 mon.a (mon.0) 216 : cluster [DBG] last_changed 2026-03-10T10:15:47.779518+0000 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791112+0000 mon.a (mon.0) 217 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791112+0000 mon.a (mon.0) 217 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791135+0000 mon.a (mon.0) 218 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:15:53.180 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791135+0000 mon.a (mon.0) 218 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791151+0000 mon.a (mon.0) 219 : cluster [DBG] election_strategy: 1 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791151+0000 mon.a (mon.0) 219 : cluster [DBG] election_strategy: 1 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791181+0000 mon.a (mon.0) 220 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791181+0000 mon.a (mon.0) 220 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791194+0000 mon.a (mon.0) 221 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791194+0000 mon.a (mon.0) 221 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791550+0000 mon.a (mon.0) 222 : cluster [DBG] fsmap 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791550+0000 mon.a (mon.0) 222 : cluster [DBG] fsmap 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791574+0000 mon.a (mon.0) 223 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791574+0000 mon.a (mon.0) 223 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791699+0000 mon.a (mon.0) 224 : cluster [DBG] mgrmap e12: a(active, since 47s) 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791699+0000 mon.a (mon.0) 224 : cluster [DBG] mgrmap e12: a(active, since 47s) 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791840+0000 mon.a (mon.0) 225 : cluster [INF] overall HEALTH_OK 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: cluster 2026-03-10T10:15:52.791840+0000 mon.a (mon.0) 225 : cluster [INF] overall HEALTH_OK 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.796216+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.796216+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.800313+0000 mon.a (mon.0) 227 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.800313+0000 mon.a (mon.0) 227 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.803506+0000 mon.a (mon.0) 228 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.803506+0000 mon.a (mon.0) 228 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.809242+0000 mon.a (mon.0) 229 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.809242+0000 mon.a (mon.0) 229 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.815771+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.815771+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.819356+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.819356+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.822535+0000 mon.a (mon.0) 232 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.822535+0000 mon.a (mon.0) 232 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.824675+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.824675+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.830599+0000 mon.a (mon.0) 234 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.181 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:52 vm00 bash[20709]: audit 2026-03-10T10:15:52.830599+0000 mon.a (mon.0) 234 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:47.765280+0000 mgr.a (mgr.14150) 49 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:47.765280+0000 mgr.a (mgr.14150) 49 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:47.782565+0000 mon.a (mon.0) 201 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:47.782565+0000 mon.a (mon.0) 201 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:47.782680+0000 mon.a (mon.0) 202 : cluster [INF] mon.a calling monitor election 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:47.782680+0000 mon.a (mon.0) 202 : cluster [INF] mon.a calling monitor election 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:47.783736+0000 mon.a (mon.0) 203 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:47.783736+0000 mon.a (mon.0) 203 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:47.908047+0000 mon.a (mon.0) 204 : audit [DBG] from='client.? 192.168.123.109:0/3793675847' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:47.908047+0000 mon.a (mon.0) 204 : audit [DBG] from='client.? 192.168.123.109:0/3793675847' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:48.777060+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:48.777060+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:49.765443+0000 mgr.a (mgr.14150) 50 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:49.765443+0000 mgr.a (mgr.14150) 50 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:49.777615+0000 mon.a (mon.0) 206 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:49.777615+0000 mon.a (mon.0) 206 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:49.784703+0000 mon.c (mon.1) 1 : cluster [INF] mon.c calling monitor election 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:49.784703+0000 mon.c (mon.1) 1 : cluster [INF] mon.c calling monitor election 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:50.521397+0000 mon.a (mon.0) 207 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:50.521397+0000 mon.a (mon.0) 207 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:50.777581+0000 mon.a (mon.0) 208 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:50.777581+0000 mon.a (mon.0) 208 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:51.521436+0000 mon.a (mon.0) 209 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:51.521436+0000 mon.a (mon.0) 209 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:51.765607+0000 mgr.a (mgr.14150) 51 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:51.765607+0000 mgr.a (mgr.14150) 51 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:51.777427+0000 mon.a (mon.0) 210 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:51.777427+0000 mon.a (mon.0) 210 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.522097+0000 mon.a (mon.0) 211 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.522097+0000 mon.a (mon.0) 211 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.777890+0000 mon.a (mon.0) 212 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.777890+0000 mon.a (mon.0) 212 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.787542+0000 mon.a (mon.0) 213 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.787542+0000 mon.a (mon.0) 213 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791023+0000 mon.a (mon.0) 214 : cluster [DBG] monmap epoch 2 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791023+0000 mon.a (mon.0) 214 : cluster [DBG] monmap epoch 2 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791049+0000 mon.a (mon.0) 215 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791049+0000 mon.a (mon.0) 215 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791062+0000 mon.a (mon.0) 216 : cluster [DBG] last_changed 2026-03-10T10:15:47.779518+0000 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791062+0000 mon.a (mon.0) 216 : cluster [DBG] last_changed 2026-03-10T10:15:47.779518+0000 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791112+0000 mon.a (mon.0) 217 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791112+0000 mon.a (mon.0) 217 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791135+0000 mon.a (mon.0) 218 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791135+0000 mon.a (mon.0) 218 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791151+0000 mon.a (mon.0) 219 : cluster [DBG] election_strategy: 1 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791151+0000 mon.a (mon.0) 219 : cluster [DBG] election_strategy: 1 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791181+0000 mon.a (mon.0) 220 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791181+0000 mon.a (mon.0) 220 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791194+0000 mon.a (mon.0) 221 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:15:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791194+0000 mon.a (mon.0) 221 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791550+0000 mon.a (mon.0) 222 : cluster [DBG] fsmap 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791550+0000 mon.a (mon.0) 222 : cluster [DBG] fsmap 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791574+0000 mon.a (mon.0) 223 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791574+0000 mon.a (mon.0) 223 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791699+0000 mon.a (mon.0) 224 : cluster [DBG] mgrmap e12: a(active, since 47s) 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791699+0000 mon.a (mon.0) 224 : cluster [DBG] mgrmap e12: a(active, since 47s) 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791840+0000 mon.a (mon.0) 225 : cluster [INF] overall HEALTH_OK 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: cluster 2026-03-10T10:15:52.791840+0000 mon.a (mon.0) 225 : cluster [INF] overall HEALTH_OK 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.796216+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.796216+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.800313+0000 mon.a (mon.0) 227 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.800313+0000 mon.a (mon.0) 227 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.803506+0000 mon.a (mon.0) 228 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.803506+0000 mon.a (mon.0) 228 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.809242+0000 mon.a (mon.0) 229 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.809242+0000 mon.a (mon.0) 229 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.815771+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.815771+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.819356+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.819356+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.822535+0000 mon.a (mon.0) 232 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.822535+0000 mon.a (mon.0) 232 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.824675+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.824675+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.830599+0000 mon.a (mon.0) 234 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:52 vm09 bash[20916]: audit 2026-03-10T10:15:52.830599+0000 mon.a (mon.0) 234 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:53.866 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-10T10:15:53.866 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph mon dump -f json 2026-03-10T10:15:54.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:52.843913+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:54.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:52.843913+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:52.848208+0000 mon.a (mon.0) 236 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:52.848208+0000 mon.a (mon.0) 236 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:52.851193+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:52.851193+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:52.851691+0000 mgr.a (mgr.14150) 52 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:52.851691+0000 mgr.a (mgr.14150) 52 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:52.851855+0000 mgr.a (mgr.14150) 53 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:52.851855+0000 mgr.a (mgr.14150) 53 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:52.853437+0000 mgr.a (mgr.14150) 54 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:52.853437+0000 mgr.a (mgr.14150) 54 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:52.926502+0000 mgr.a (mgr.14150) 55 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:52.926502+0000 mgr.a (mgr.14150) 55 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:52.926582+0000 mgr.a (mgr.14150) 56 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:52.926582+0000 mgr.a (mgr.14150) 56 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:52.932743+0000 mgr.a (mgr.14150) 57 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:52.932743+0000 mgr.a (mgr.14150) 57 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.003043+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.003043+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.006713+0000 mon.a (mon.0) 239 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.006713+0000 mon.a (mon.0) 239 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.009826+0000 mon.a (mon.0) 240 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.009826+0000 mon.a (mon.0) 240 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.012640+0000 mon.a (mon.0) 241 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.012640+0000 mon.a (mon.0) 241 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.015453+0000 mon.a (mon.0) 242 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.015453+0000 mon.a (mon.0) 242 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.018444+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.018444+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.020920+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.020920+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:53.021215+0000 mgr.a (mgr.14150) 58 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:53.021215+0000 mgr.a (mgr.14150) 58 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:53.035147+0000 mgr.a (mgr.14150) 59 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:53.035147+0000 mgr.a (mgr.14150) 59 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.035396+0000 mon.a (mon.0) 245 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.035396+0000 mon.a (mon.0) 245 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.035817+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.035817+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.036164+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.036164+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:53.036563+0000 mgr.a (mgr.14150) 60 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:53.036563+0000 mgr.a (mgr.14150) 60 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.522113+0000 mon.a (mon.0) 248 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.522113+0000 mon.a (mon.0) 248 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.575804+0000 mon.a (mon.0) 249 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.575804+0000 mon.a (mon.0) 249 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.579152+0000 mon.a (mon.0) 250 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.579152+0000 mon.a (mon.0) 250 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:53.579862+0000 mgr.a (mgr.14150) 61 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:53.579862+0000 mgr.a (mgr.14150) 61 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.580020+0000 mon.a (mon.0) 251 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.580020+0000 mon.a (mon.0) 251 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.580482+0000 mon.a (mon.0) 252 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.580482+0000 mon.a (mon.0) 252 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.580860+0000 mon.a (mon.0) 253 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.580860+0000 mon.a (mon.0) 253 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:53.581298+0000 mgr.a (mgr.14150) 62 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: cephadm 2026-03-10T10:15:53.581298+0000 mgr.a (mgr.14150) 62 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:15:54.169 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.777694+0000 mon.a (mon.0) 254 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:54.170 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:53 vm00 bash[20709]: audit 2026-03-10T10:15:53.777694+0000 mon.a (mon.0) 254 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:54.254 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:52.843913+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:54.254 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:52.843913+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:52.848208+0000 mon.a (mon.0) 236 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:52.848208+0000 mon.a (mon.0) 236 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:52.851193+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:52.851193+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:52.851691+0000 mgr.a (mgr.14150) 52 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:52.851691+0000 mgr.a (mgr.14150) 52 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:52.851855+0000 mgr.a (mgr.14150) 53 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:52.851855+0000 mgr.a (mgr.14150) 53 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:52.853437+0000 mgr.a (mgr.14150) 54 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:52.853437+0000 mgr.a (mgr.14150) 54 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:52.926502+0000 mgr.a (mgr.14150) 55 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:52.926502+0000 mgr.a (mgr.14150) 55 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:52.926582+0000 mgr.a (mgr.14150) 56 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:52.926582+0000 mgr.a (mgr.14150) 56 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:52.932743+0000 mgr.a (mgr.14150) 57 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:52.932743+0000 mgr.a (mgr.14150) 57 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.003043+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.003043+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.006713+0000 mon.a (mon.0) 239 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.006713+0000 mon.a (mon.0) 239 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.009826+0000 mon.a (mon.0) 240 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.009826+0000 mon.a (mon.0) 240 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.012640+0000 mon.a (mon.0) 241 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.012640+0000 mon.a (mon.0) 241 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.015453+0000 mon.a (mon.0) 242 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.015453+0000 mon.a (mon.0) 242 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.018444+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.018444+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.020920+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.020920+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:53.021215+0000 mgr.a (mgr.14150) 58 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:53.021215+0000 mgr.a (mgr.14150) 58 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:53.035147+0000 mgr.a (mgr.14150) 59 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:53.035147+0000 mgr.a (mgr.14150) 59 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.035396+0000 mon.a (mon.0) 245 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.035396+0000 mon.a (mon.0) 245 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.035817+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:15:54.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.035817+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.036164+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.036164+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:53.036563+0000 mgr.a (mgr.14150) 60 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:53.036563+0000 mgr.a (mgr.14150) 60 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.522113+0000 mon.a (mon.0) 248 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.522113+0000 mon.a (mon.0) 248 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.575804+0000 mon.a (mon.0) 249 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.575804+0000 mon.a (mon.0) 249 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.579152+0000 mon.a (mon.0) 250 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.579152+0000 mon.a (mon.0) 250 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:53.579862+0000 mgr.a (mgr.14150) 61 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:53.579862+0000 mgr.a (mgr.14150) 61 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.580020+0000 mon.a (mon.0) 251 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.580020+0000 mon.a (mon.0) 251 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.580482+0000 mon.a (mon.0) 252 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.580482+0000 mon.a (mon.0) 252 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.580860+0000 mon.a (mon.0) 253 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.580860+0000 mon.a (mon.0) 253 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:53.581298+0000 mgr.a (mgr.14150) 62 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: cephadm 2026-03-10T10:15:53.581298+0000 mgr.a (mgr.14150) 62 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.777694+0000 mon.a (mon.0) 254 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:54.256 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:53 vm09 bash[20916]: audit 2026-03-10T10:15:53.777694+0000 mon.a (mon.0) 254 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:47.765280+0000 mgr.a (mgr.14150) 49 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:47.765280+0000 mgr.a (mgr.14150) 49 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:47.782565+0000 mon.a (mon.0) 201 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:47.782565+0000 mon.a (mon.0) 201 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:47.782680+0000 mon.a (mon.0) 202 : cluster [INF] mon.a calling monitor election 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:47.782680+0000 mon.a (mon.0) 202 : cluster [INF] mon.a calling monitor election 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:47.783736+0000 mon.a (mon.0) 203 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:47.783736+0000 mon.a (mon.0) 203 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:47.908047+0000 mon.a (mon.0) 204 : audit [DBG] from='client.? 192.168.123.109:0/3793675847' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:47.908047+0000 mon.a (mon.0) 204 : audit [DBG] from='client.? 192.168.123.109:0/3793675847' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:48.777060+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:48.777060+0000 mon.a (mon.0) 205 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:49.765443+0000 mgr.a (mgr.14150) 50 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:49.765443+0000 mgr.a (mgr.14150) 50 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:49.777615+0000 mon.a (mon.0) 206 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:49.777615+0000 mon.a (mon.0) 206 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:49.784703+0000 mon.c (mon.1) 1 : cluster [INF] mon.c calling monitor election 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:49.784703+0000 mon.c (mon.1) 1 : cluster [INF] mon.c calling monitor election 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:50.521397+0000 mon.a (mon.0) 207 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:50.521397+0000 mon.a (mon.0) 207 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:50.777581+0000 mon.a (mon.0) 208 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:50.777581+0000 mon.a (mon.0) 208 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:51.521436+0000 mon.a (mon.0) 209 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:51.521436+0000 mon.a (mon.0) 209 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:51.765607+0000 mgr.a (mgr.14150) 51 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:51.765607+0000 mgr.a (mgr.14150) 51 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:51.777427+0000 mon.a (mon.0) 210 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:51.777427+0000 mon.a (mon.0) 210 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.522097+0000 mon.a (mon.0) 211 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:54.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.522097+0000 mon.a (mon.0) 211 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.777890+0000 mon.a (mon.0) 212 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.777890+0000 mon.a (mon.0) 212 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.787542+0000 mon.a (mon.0) 213 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.787542+0000 mon.a (mon.0) 213 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791023+0000 mon.a (mon.0) 214 : cluster [DBG] monmap epoch 2 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791023+0000 mon.a (mon.0) 214 : cluster [DBG] monmap epoch 2 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791049+0000 mon.a (mon.0) 215 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791049+0000 mon.a (mon.0) 215 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791062+0000 mon.a (mon.0) 216 : cluster [DBG] last_changed 2026-03-10T10:15:47.779518+0000 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791062+0000 mon.a (mon.0) 216 : cluster [DBG] last_changed 2026-03-10T10:15:47.779518+0000 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791112+0000 mon.a (mon.0) 217 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791112+0000 mon.a (mon.0) 217 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791135+0000 mon.a (mon.0) 218 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791135+0000 mon.a (mon.0) 218 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791151+0000 mon.a (mon.0) 219 : cluster [DBG] election_strategy: 1 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791151+0000 mon.a (mon.0) 219 : cluster [DBG] election_strategy: 1 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791181+0000 mon.a (mon.0) 220 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791181+0000 mon.a (mon.0) 220 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791194+0000 mon.a (mon.0) 221 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791194+0000 mon.a (mon.0) 221 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791550+0000 mon.a (mon.0) 222 : cluster [DBG] fsmap 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791550+0000 mon.a (mon.0) 222 : cluster [DBG] fsmap 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791574+0000 mon.a (mon.0) 223 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791574+0000 mon.a (mon.0) 223 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791699+0000 mon.a (mon.0) 224 : cluster [DBG] mgrmap e12: a(active, since 47s) 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791699+0000 mon.a (mon.0) 224 : cluster [DBG] mgrmap e12: a(active, since 47s) 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791840+0000 mon.a (mon.0) 225 : cluster [INF] overall HEALTH_OK 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cluster 2026-03-10T10:15:52.791840+0000 mon.a (mon.0) 225 : cluster [INF] overall HEALTH_OK 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.796216+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.796216+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.800313+0000 mon.a (mon.0) 227 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.800313+0000 mon.a (mon.0) 227 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.803506+0000 mon.a (mon.0) 228 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.803506+0000 mon.a (mon.0) 228 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.809242+0000 mon.a (mon.0) 229 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.833 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.809242+0000 mon.a (mon.0) 229 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.815771+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.815771+0000 mon.a (mon.0) 230 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.819356+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.819356+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.822535+0000 mon.a (mon.0) 232 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.822535+0000 mon.a (mon.0) 232 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.824675+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.824675+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.830599+0000 mon.a (mon.0) 234 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.830599+0000 mon.a (mon.0) 234 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.843913+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.843913+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.848208+0000 mon.a (mon.0) 236 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.848208+0000 mon.a (mon.0) 236 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.851193+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:52.851193+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:52.851691+0000 mgr.a (mgr.14150) 52 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:52.851691+0000 mgr.a (mgr.14150) 52 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:52.851855+0000 mgr.a (mgr.14150) 53 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:52.851855+0000 mgr.a (mgr.14150) 53 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:52.853437+0000 mgr.a (mgr.14150) 54 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:52.853437+0000 mgr.a (mgr.14150) 54 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:52.926502+0000 mgr.a (mgr.14150) 55 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:52.926502+0000 mgr.a (mgr.14150) 55 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:52.926582+0000 mgr.a (mgr.14150) 56 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:52.926582+0000 mgr.a (mgr.14150) 56 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:52.932743+0000 mgr.a (mgr.14150) 57 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:52.932743+0000 mgr.a (mgr.14150) 57 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.003043+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.003043+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.006713+0000 mon.a (mon.0) 239 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.006713+0000 mon.a (mon.0) 239 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.009826+0000 mon.a (mon.0) 240 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.009826+0000 mon.a (mon.0) 240 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.012640+0000 mon.a (mon.0) 241 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.012640+0000 mon.a (mon.0) 241 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.015453+0000 mon.a (mon.0) 242 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.015453+0000 mon.a (mon.0) 242 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.018444+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.018444+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.020920+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.020920+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:53.021215+0000 mgr.a (mgr.14150) 58 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:53.021215+0000 mgr.a (mgr.14150) 58 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:53.035147+0000 mgr.a (mgr.14150) 59 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:53.035147+0000 mgr.a (mgr.14150) 59 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.035396+0000 mon.a (mon.0) 245 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.035396+0000 mon.a (mon.0) 245 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:54.834 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.035817+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.035817+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.036164+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.036164+0000 mon.a (mon.0) 247 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:53.036563+0000 mgr.a (mgr.14150) 60 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:53.036563+0000 mgr.a (mgr.14150) 60 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.522113+0000 mon.a (mon.0) 248 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.522113+0000 mon.a (mon.0) 248 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.575804+0000 mon.a (mon.0) 249 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.575804+0000 mon.a (mon.0) 249 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.579152+0000 mon.a (mon.0) 250 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.579152+0000 mon.a (mon.0) 250 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:53.579862+0000 mgr.a (mgr.14150) 61 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:53.579862+0000 mgr.a (mgr.14150) 61 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.580020+0000 mon.a (mon.0) 251 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.580020+0000 mon.a (mon.0) 251 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.580482+0000 mon.a (mon.0) 252 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.580482+0000 mon.a (mon.0) 252 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.580860+0000 mon.a (mon.0) 253 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.580860+0000 mon.a (mon.0) 253 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:53.581298+0000 mgr.a (mgr.14150) 62 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: cephadm 2026-03-10T10:15:53.581298+0000 mgr.a (mgr.14150) 62 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.777694+0000 mon.a (mon.0) 254 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:54.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:54 vm03 bash[21274]: audit 2026-03-10T10:15:53.777694+0000 mon.a (mon.0) 254 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:55.168 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:15:54 vm00 bash[20974]: debug 2026-03-10T10:15:54.773+0000 7f26d7a3b640 -1 mgr.server handle_report got status from non-daemon mon.c 2026-03-10T10:15:58.042 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.c/config 2026-03-10T10:15:59.733 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 3 2026-03-10T10:15:59.733 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:15:59.733 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":3,"fsid":"9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad","modified":"2026-03-10T10:15:54.531145Z","created":"2026-03-10T10:14:44.248116Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:3300","nonce":0},{"type":"v1","addr":"192.168.123.100:6789","nonce":0}]},"addr":"192.168.123.100:6789/0","public_addr":"192.168.123.100:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:3300","nonce":0},{"type":"v1","addr":"192.168.123.109:6789","nonce":0}]},"addr":"192.168.123.109:6789/0","public_addr":"192.168.123.109:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":2,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:3300","nonce":0},{"type":"v1","addr":"192.168.123.103:6789","nonce":0}]},"addr":"192.168.123.103:6789/0","public_addr":"192.168.123.103:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1,2]} 2026-03-10T10:15:59.782 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-10T10:15:59.782 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph config generate-minimal-conf 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:53.765767+0000 mgr.a (mgr.14150) 63 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:53.765767+0000 mgr.a (mgr.14150) 63 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cephadm 2026-03-10T10:15:53.978588+0000 mgr.a (mgr.14150) 64 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cephadm 2026-03-10T10:15:53.978588+0000 mgr.a (mgr.14150) 64 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:54.535196+0000 mon.a (mon.0) 262 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:54.535196+0000 mon.a (mon.0) 262 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:54.535291+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:54.535291+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:54.535326+0000 mon.a (mon.0) 264 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:54.535326+0000 mon.a (mon.0) 264 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:54.535441+0000 mon.a (mon.0) 265 : cluster [INF] mon.a calling monitor election 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:54.535441+0000 mon.a (mon.0) 265 : cluster [INF] mon.a calling monitor election 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:54.542841+0000 mon.c (mon.1) 2 : cluster [INF] mon.c calling monitor election 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:54.542841+0000 mon.c (mon.1) 2 : cluster [INF] mon.c calling monitor election 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:55.521950+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:55.521950+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:55.765942+0000 mgr.a (mgr.14150) 65 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:55.765942+0000 mgr.a (mgr.14150) 65 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:56.522272+0000 mon.a (mon.0) 267 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:56.522272+0000 mon.a (mon.0) 267 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:56.532807+0000 mon.b (mon.2) 1 : cluster [INF] mon.b calling monitor election 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:56.532807+0000 mon.b (mon.2) 1 : cluster [INF] mon.b calling monitor election 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:57.522340+0000 mon.a (mon.0) 268 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:57.522340+0000 mon.a (mon.0) 268 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.819 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:57.766425+0000 mgr.a (mgr.14150) 66 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:57.766425+0000 mgr.a (mgr.14150) 66 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:58.522483+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:58.522483+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:59.522417+0000 mon.a (mon.0) 270 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:59.522417+0000 mon.a (mon.0) 270 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.539489+0000 mon.a (mon.0) 271 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.539489+0000 mon.a (mon.0) 271 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542742+0000 mon.a (mon.0) 272 : cluster [DBG] monmap epoch 3 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542742+0000 mon.a (mon.0) 272 : cluster [DBG] monmap epoch 3 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542748+0000 mon.a (mon.0) 273 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542748+0000 mon.a (mon.0) 273 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542779+0000 mon.a (mon.0) 274 : cluster [DBG] last_changed 2026-03-10T10:15:54.531145+0000 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542779+0000 mon.a (mon.0) 274 : cluster [DBG] last_changed 2026-03-10T10:15:54.531145+0000 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542782+0000 mon.a (mon.0) 275 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542782+0000 mon.a (mon.0) 275 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542786+0000 mon.a (mon.0) 276 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542786+0000 mon.a (mon.0) 276 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542789+0000 mon.a (mon.0) 277 : cluster [DBG] election_strategy: 1 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542789+0000 mon.a (mon.0) 277 : cluster [DBG] election_strategy: 1 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542793+0000 mon.a (mon.0) 278 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542793+0000 mon.a (mon.0) 278 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542795+0000 mon.a (mon.0) 279 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542795+0000 mon.a (mon.0) 279 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542799+0000 mon.a (mon.0) 280 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.542799+0000 mon.a (mon.0) 280 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.543030+0000 mon.a (mon.0) 281 : cluster [DBG] fsmap 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.543030+0000 mon.a (mon.0) 281 : cluster [DBG] fsmap 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.543041+0000 mon.a (mon.0) 282 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.543041+0000 mon.a (mon.0) 282 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.543153+0000 mon.a (mon.0) 283 : cluster [DBG] mgrmap e12: a(active, since 53s) 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.543153+0000 mon.a (mon.0) 283 : cluster [DBG] mgrmap e12: a(active, since 53s) 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.543216+0000 mon.a (mon.0) 284 : cluster [INF] overall HEALTH_OK 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: cluster 2026-03-10T10:15:59.543216+0000 mon.a (mon.0) 284 : cluster [INF] overall HEALTH_OK 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:59.545911+0000 mon.a (mon.0) 285 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:59.545911+0000 mon.a (mon.0) 285 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:59.549110+0000 mon.a (mon.0) 286 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:59.549110+0000 mon.a (mon.0) 286 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:59.551823+0000 mon.a (mon.0) 287 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:59.551823+0000 mon.a (mon.0) 287 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:59.554392+0000 mon.a (mon.0) 288 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:59.554392+0000 mon.a (mon.0) 288 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:59.557208+0000 mon.a (mon.0) 289 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:59.557208+0000 mon.a (mon.0) 289 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:59.560406+0000 mon.a (mon.0) 290 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.820 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:15:59 vm03 bash[21274]: audit 2026-03-10T10:15:59.560406+0000 mon.a (mon.0) 290 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:53.765767+0000 mgr.a (mgr.14150) 63 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:53.765767+0000 mgr.a (mgr.14150) 63 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cephadm 2026-03-10T10:15:53.978588+0000 mgr.a (mgr.14150) 64 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cephadm 2026-03-10T10:15:53.978588+0000 mgr.a (mgr.14150) 64 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:54.535196+0000 mon.a (mon.0) 262 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:54.535196+0000 mon.a (mon.0) 262 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:54.535291+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:54.535291+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:54.535326+0000 mon.a (mon.0) 264 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:54.535326+0000 mon.a (mon.0) 264 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:54.535441+0000 mon.a (mon.0) 265 : cluster [INF] mon.a calling monitor election 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:54.535441+0000 mon.a (mon.0) 265 : cluster [INF] mon.a calling monitor election 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:54.542841+0000 mon.c (mon.1) 2 : cluster [INF] mon.c calling monitor election 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:54.542841+0000 mon.c (mon.1) 2 : cluster [INF] mon.c calling monitor election 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:55.521950+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:55.521950+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:55.765942+0000 mgr.a (mgr.14150) 65 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:55.765942+0000 mgr.a (mgr.14150) 65 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:56.522272+0000 mon.a (mon.0) 267 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:56.522272+0000 mon.a (mon.0) 267 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:56.532807+0000 mon.b (mon.2) 1 : cluster [INF] mon.b calling monitor election 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:56.532807+0000 mon.b (mon.2) 1 : cluster [INF] mon.b calling monitor election 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:57.522340+0000 mon.a (mon.0) 268 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:57.522340+0000 mon.a (mon.0) 268 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:57.766425+0000 mgr.a (mgr.14150) 66 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:57.766425+0000 mgr.a (mgr.14150) 66 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:58.522483+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:58.522483+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:59.522417+0000 mon.a (mon.0) 270 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:59.522417+0000 mon.a (mon.0) 270 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.539489+0000 mon.a (mon.0) 271 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.539489+0000 mon.a (mon.0) 271 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542742+0000 mon.a (mon.0) 272 : cluster [DBG] monmap epoch 3 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542742+0000 mon.a (mon.0) 272 : cluster [DBG] monmap epoch 3 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542748+0000 mon.a (mon.0) 273 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542748+0000 mon.a (mon.0) 273 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542779+0000 mon.a (mon.0) 274 : cluster [DBG] last_changed 2026-03-10T10:15:54.531145+0000 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542779+0000 mon.a (mon.0) 274 : cluster [DBG] last_changed 2026-03-10T10:15:54.531145+0000 2026-03-10T10:15:59.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542782+0000 mon.a (mon.0) 275 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542782+0000 mon.a (mon.0) 275 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542786+0000 mon.a (mon.0) 276 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542786+0000 mon.a (mon.0) 276 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542789+0000 mon.a (mon.0) 277 : cluster [DBG] election_strategy: 1 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542789+0000 mon.a (mon.0) 277 : cluster [DBG] election_strategy: 1 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542793+0000 mon.a (mon.0) 278 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542793+0000 mon.a (mon.0) 278 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542795+0000 mon.a (mon.0) 279 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542795+0000 mon.a (mon.0) 279 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542799+0000 mon.a (mon.0) 280 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.542799+0000 mon.a (mon.0) 280 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.543030+0000 mon.a (mon.0) 281 : cluster [DBG] fsmap 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.543030+0000 mon.a (mon.0) 281 : cluster [DBG] fsmap 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.543041+0000 mon.a (mon.0) 282 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.543041+0000 mon.a (mon.0) 282 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.543153+0000 mon.a (mon.0) 283 : cluster [DBG] mgrmap e12: a(active, since 53s) 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.543153+0000 mon.a (mon.0) 283 : cluster [DBG] mgrmap e12: a(active, since 53s) 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.543216+0000 mon.a (mon.0) 284 : cluster [INF] overall HEALTH_OK 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: cluster 2026-03-10T10:15:59.543216+0000 mon.a (mon.0) 284 : cluster [INF] overall HEALTH_OK 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:59.545911+0000 mon.a (mon.0) 285 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:59.545911+0000 mon.a (mon.0) 285 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:59.549110+0000 mon.a (mon.0) 286 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:59.549110+0000 mon.a (mon.0) 286 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:59.551823+0000 mon.a (mon.0) 287 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:59.551823+0000 mon.a (mon.0) 287 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:59.554392+0000 mon.a (mon.0) 288 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:59.554392+0000 mon.a (mon.0) 288 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:59.557208+0000 mon.a (mon.0) 289 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:59.557208+0000 mon.a (mon.0) 289 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:59.560406+0000 mon.a (mon.0) 290 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:15:59.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:15:59 vm00 bash[20709]: audit 2026-03-10T10:15:59.560406+0000 mon.a (mon.0) 290 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:53.765767+0000 mgr.a (mgr.14150) 63 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:00.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:53.765767+0000 mgr.a (mgr.14150) 63 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:00.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cephadm 2026-03-10T10:15:53.978588+0000 mgr.a (mgr.14150) 64 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:16:00.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cephadm 2026-03-10T10:15:53.978588+0000 mgr.a (mgr.14150) 64 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:16:00.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:54.535196+0000 mon.a (mon.0) 262 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:16:00.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:54.535196+0000 mon.a (mon.0) 262 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:16:00.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:54.535291+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:00.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:54.535291+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:00.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:54.535326+0000 mon.a (mon.0) 264 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:16:00.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:54.535326+0000 mon.a (mon.0) 264 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:16:00.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:54.535441+0000 mon.a (mon.0) 265 : cluster [INF] mon.a calling monitor election 2026-03-10T10:16:00.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:54.535441+0000 mon.a (mon.0) 265 : cluster [INF] mon.a calling monitor election 2026-03-10T10:16:00.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:54.542841+0000 mon.c (mon.1) 2 : cluster [INF] mon.c calling monitor election 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:54.542841+0000 mon.c (mon.1) 2 : cluster [INF] mon.c calling monitor election 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:55.521950+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:55.521950+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:55.765942+0000 mgr.a (mgr.14150) 65 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:55.765942+0000 mgr.a (mgr.14150) 65 : cluster [DBG] pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:56.522272+0000 mon.a (mon.0) 267 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:56.522272+0000 mon.a (mon.0) 267 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:56.532807+0000 mon.b (mon.2) 1 : cluster [INF] mon.b calling monitor election 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:56.532807+0000 mon.b (mon.2) 1 : cluster [INF] mon.b calling monitor election 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:57.522340+0000 mon.a (mon.0) 268 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:57.522340+0000 mon.a (mon.0) 268 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:57.766425+0000 mgr.a (mgr.14150) 66 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:57.766425+0000 mgr.a (mgr.14150) 66 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:58.522483+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:58.522483+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:59.522417+0000 mon.a (mon.0) 270 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:59.522417+0000 mon.a (mon.0) 270 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.539489+0000 mon.a (mon.0) 271 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.539489+0000 mon.a (mon.0) 271 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542742+0000 mon.a (mon.0) 272 : cluster [DBG] monmap epoch 3 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542742+0000 mon.a (mon.0) 272 : cluster [DBG] monmap epoch 3 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542748+0000 mon.a (mon.0) 273 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542748+0000 mon.a (mon.0) 273 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542779+0000 mon.a (mon.0) 274 : cluster [DBG] last_changed 2026-03-10T10:15:54.531145+0000 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542779+0000 mon.a (mon.0) 274 : cluster [DBG] last_changed 2026-03-10T10:15:54.531145+0000 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542782+0000 mon.a (mon.0) 275 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542782+0000 mon.a (mon.0) 275 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542786+0000 mon.a (mon.0) 276 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542786+0000 mon.a (mon.0) 276 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542789+0000 mon.a (mon.0) 277 : cluster [DBG] election_strategy: 1 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542789+0000 mon.a (mon.0) 277 : cluster [DBG] election_strategy: 1 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542793+0000 mon.a (mon.0) 278 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:16:00.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542793+0000 mon.a (mon.0) 278 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542795+0000 mon.a (mon.0) 279 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542795+0000 mon.a (mon.0) 279 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542799+0000 mon.a (mon.0) 280 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.542799+0000 mon.a (mon.0) 280 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.543030+0000 mon.a (mon.0) 281 : cluster [DBG] fsmap 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.543030+0000 mon.a (mon.0) 281 : cluster [DBG] fsmap 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.543041+0000 mon.a (mon.0) 282 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.543041+0000 mon.a (mon.0) 282 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.543153+0000 mon.a (mon.0) 283 : cluster [DBG] mgrmap e12: a(active, since 53s) 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.543153+0000 mon.a (mon.0) 283 : cluster [DBG] mgrmap e12: a(active, since 53s) 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.543216+0000 mon.a (mon.0) 284 : cluster [INF] overall HEALTH_OK 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: cluster 2026-03-10T10:15:59.543216+0000 mon.a (mon.0) 284 : cluster [INF] overall HEALTH_OK 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:59.545911+0000 mon.a (mon.0) 285 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:59.545911+0000 mon.a (mon.0) 285 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:59.549110+0000 mon.a (mon.0) 286 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:59.549110+0000 mon.a (mon.0) 286 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:59.551823+0000 mon.a (mon.0) 287 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:59.551823+0000 mon.a (mon.0) 287 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:59.554392+0000 mon.a (mon.0) 288 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:59.554392+0000 mon.a (mon.0) 288 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:59.557208+0000 mon.a (mon.0) 289 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:59.557208+0000 mon.a (mon.0) 289 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:59.560406+0000 mon.a (mon.0) 290 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:15:59 vm09 bash[20916]: audit 2026-03-10T10:15:59.560406+0000 mon.a (mon.0) 290 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:15:59.731836+0000 mon.a (mon.0) 291 : audit [DBG] from='client.? 192.168.123.109:0/3701825091' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:16:00.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:15:59.731836+0000 mon.a (mon.0) 291 : audit [DBG] from='client.? 192.168.123.109:0/3701825091' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:16:00.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:15:59.899900+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:15:59.899900+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:15:59.903113+0000 mon.a (mon.0) 293 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:15:59.903113+0000 mon.a (mon.0) 293 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:15:59.906581+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:15:59.906581+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.211657+0000 mon.a (mon.0) 295 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.211657+0000 mon.a (mon.0) 295 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.215101+0000 mon.a (mon.0) 296 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.215101+0000 mon.a (mon.0) 296 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.216559+0000 mon.a (mon.0) 297 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.216559+0000 mon.a (mon.0) 297 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.217738+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.217738+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.218216+0000 mon.a (mon.0) 299 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.218216+0000 mon.a (mon.0) 299 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.222997+0000 mon.a (mon.0) 300 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.222997+0000 mon.a (mon.0) 300 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.308784+0000 mon.a (mon.0) 301 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.308784+0000 mon.a (mon.0) 301 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.311809+0000 mon.a (mon.0) 302 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.311809+0000 mon.a (mon.0) 302 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.314445+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.314445+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.317274+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.317274+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.319913+0000 mon.a (mon.0) 305 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.319913+0000 mon.a (mon.0) 305 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.322576+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.322576+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.325339+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.325339+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.340066+0000 mon.a (mon.0) 308 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.340066+0000 mon.a (mon.0) 308 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.340518+0000 mon.a (mon.0) 309 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.340518+0000 mon.a (mon.0) 309 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.340879+0000 mon.a (mon.0) 310 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.340879+0000 mon.a (mon.0) 310 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.522667+0000 mon.a (mon.0) 311 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:00.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:00 vm00 bash[20709]: audit 2026-03-10T10:16:00.522667+0000 mon.a (mon.0) 311 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:00.999 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:15:59.731836+0000 mon.a (mon.0) 291 : audit [DBG] from='client.? 192.168.123.109:0/3701825091' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:16:00.999 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:15:59.731836+0000 mon.a (mon.0) 291 : audit [DBG] from='client.? 192.168.123.109:0/3701825091' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:16:00.999 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:15:59.899900+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.999 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:15:59.899900+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.999 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:15:59.903113+0000 mon.a (mon.0) 293 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.999 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:15:59.903113+0000 mon.a (mon.0) 293 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.999 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:15:59.906581+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.999 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:15:59.906581+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.999 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.211657+0000 mon.a (mon.0) 295 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.999 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.211657+0000 mon.a (mon.0) 295 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.999 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.215101+0000 mon.a (mon.0) 296 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.999 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.215101+0000 mon.a (mon.0) 296 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:00.999 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.216559+0000 mon.a (mon.0) 297 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:00.999 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.216559+0000 mon.a (mon.0) 297 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:00.999 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.217738+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.217738+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.218216+0000 mon.a (mon.0) 299 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.218216+0000 mon.a (mon.0) 299 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.222997+0000 mon.a (mon.0) 300 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.222997+0000 mon.a (mon.0) 300 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.308784+0000 mon.a (mon.0) 301 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.308784+0000 mon.a (mon.0) 301 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.311809+0000 mon.a (mon.0) 302 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.311809+0000 mon.a (mon.0) 302 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.314445+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.314445+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.317274+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.317274+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.319913+0000 mon.a (mon.0) 305 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.319913+0000 mon.a (mon.0) 305 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.322576+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.322576+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.325339+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.325339+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.340066+0000 mon.a (mon.0) 308 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.340066+0000 mon.a (mon.0) 308 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.340518+0000 mon.a (mon.0) 309 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.340518+0000 mon.a (mon.0) 309 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.340879+0000 mon.a (mon.0) 310 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.340879+0000 mon.a (mon.0) 310 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.522667+0000 mon.a (mon.0) 311 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:01.000 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:00 vm03 bash[21274]: audit 2026-03-10T10:16:00.522667+0000 mon.a (mon.0) 311 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:01.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:15:59.731836+0000 mon.a (mon.0) 291 : audit [DBG] from='client.? 192.168.123.109:0/3701825091' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:16:01.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:15:59.731836+0000 mon.a (mon.0) 291 : audit [DBG] from='client.? 192.168.123.109:0/3701825091' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:16:01.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:15:59.899900+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:15:59.899900+0000 mon.a (mon.0) 292 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:15:59.903113+0000 mon.a (mon.0) 293 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:15:59.903113+0000 mon.a (mon.0) 293 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:15:59.906581+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:15:59.906581+0000 mon.a (mon.0) 294 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.211657+0000 mon.a (mon.0) 295 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.211657+0000 mon.a (mon.0) 295 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.215101+0000 mon.a (mon.0) 296 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.215101+0000 mon.a (mon.0) 296 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.216559+0000 mon.a (mon.0) 297 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.216559+0000 mon.a (mon.0) 297 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.217738+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.217738+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.218216+0000 mon.a (mon.0) 299 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.218216+0000 mon.a (mon.0) 299 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.222997+0000 mon.a (mon.0) 300 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.222997+0000 mon.a (mon.0) 300 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.308784+0000 mon.a (mon.0) 301 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.308784+0000 mon.a (mon.0) 301 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.311809+0000 mon.a (mon.0) 302 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.311809+0000 mon.a (mon.0) 302 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.314445+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.314445+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.317274+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.317274+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.319913+0000 mon.a (mon.0) 305 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.319913+0000 mon.a (mon.0) 305 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.322576+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.322576+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.325339+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.325339+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.340066+0000 mon.a (mon.0) 308 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.340066+0000 mon.a (mon.0) 308 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.340518+0000 mon.a (mon.0) 309 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.340518+0000 mon.a (mon.0) 309 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.340879+0000 mon.a (mon.0) 310 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.340879+0000 mon.a (mon.0) 310 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.522667+0000 mon.a (mon.0) 311 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:01.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:00 vm09 bash[20916]: audit 2026-03-10T10:16:00.522667+0000 mon.a (mon.0) 311 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:16:01.632 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cluster 2026-03-10T10:15:59.766573+0000 mgr.a (mgr.14150) 67 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:01.632 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cluster 2026-03-10T10:15:59.766573+0000 mgr.a (mgr.14150) 67 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:01.632 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.218832+0000 mgr.a (mgr.14150) 68 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:16:01.632 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.218832+0000 mgr.a (mgr.14150) 68 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:16:01.632 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.218921+0000 mgr.a (mgr.14150) 69 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:16:01.632 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.218921+0000 mgr.a (mgr.14150) 69 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:16:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cluster 2026-03-10T10:15:59.766573+0000 mgr.a (mgr.14150) 67 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cluster 2026-03-10T10:15:59.766573+0000 mgr.a (mgr.14150) 67 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.218832+0000 mgr.a (mgr.14150) 68 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:16:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.218832+0000 mgr.a (mgr.14150) 68 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:16:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.218921+0000 mgr.a (mgr.14150) 69 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:16:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.218921+0000 mgr.a (mgr.14150) 69 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:16:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.218970+0000 mgr.a (mgr.14150) 70 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:16:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.218970+0000 mgr.a (mgr.14150) 70 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:16:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.263858+0000 mgr.a (mgr.14150) 71 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.263858+0000 mgr.a (mgr.14150) 71 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.264184+0000 mgr.a (mgr.14150) 72 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:01.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.264184+0000 mgr.a (mgr.14150) 72 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.264288+0000 mgr.a (mgr.14150) 73 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.264288+0000 mgr.a (mgr.14150) 73 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.339843+0000 mgr.a (mgr.14150) 74 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.339843+0000 mgr.a (mgr.14150) 74 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.341422+0000 mgr.a (mgr.14150) 75 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.341422+0000 mgr.a (mgr.14150) 75 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:00.708154+0000 mon.a (mon.0) 312 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:00.708154+0000 mon.a (mon.0) 312 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:00.712272+0000 mon.a (mon.0) 313 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:00.712272+0000 mon.a (mon.0) 313 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:00.713204+0000 mon.a (mon.0) 314 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:00.713204+0000 mon.a (mon.0) 314 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:00.713706+0000 mon.a (mon.0) 315 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:00.713706+0000 mon.a (mon.0) 315 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:00.714113+0000 mon.a (mon.0) 316 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:00.714113+0000 mon.a (mon.0) 316 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:01.072905+0000 mon.a (mon.0) 317 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:01.072905+0000 mon.a (mon.0) 317 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:01.077354+0000 mon.a (mon.0) 318 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:01.077354+0000 mon.a (mon.0) 318 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:01.078602+0000 mon.a (mon.0) 319 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:01.078602+0000 mon.a (mon.0) 319 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:01.078984+0000 mon.a (mon.0) 320 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:01.078984+0000 mon.a (mon.0) 320 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:01.079347+0000 mon.a (mon.0) 321 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:01.079347+0000 mon.a (mon.0) 321 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:01.425733+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:01.425733+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:01.429484+0000 mon.a (mon.0) 323 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20709]: audit 2026-03-10T10:16:01.429484+0000 mon.a (mon.0) 323 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:01.919 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:16:01 vm00 bash[20974]: debug 2026-03-10T10:16:01.521+0000 7f26d7a3b640 -1 mgr.server handle_report got status from non-daemon mon.b 2026-03-10T10:16:02.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.218970+0000 mgr.a (mgr.14150) 70 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:16:02.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.218970+0000 mgr.a (mgr.14150) 70 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:16:02.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.263858+0000 mgr.a (mgr.14150) 71 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:02.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.263858+0000 mgr.a (mgr.14150) 71 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:02.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.264184+0000 mgr.a (mgr.14150) 72 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:02.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.264184+0000 mgr.a (mgr.14150) 72 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.264288+0000 mgr.a (mgr.14150) 73 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.264288+0000 mgr.a (mgr.14150) 73 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.339843+0000 mgr.a (mgr.14150) 74 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.339843+0000 mgr.a (mgr.14150) 74 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.341422+0000 mgr.a (mgr.14150) 75 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.341422+0000 mgr.a (mgr.14150) 75 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:00.708154+0000 mon.a (mon.0) 312 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:00.708154+0000 mon.a (mon.0) 312 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:00.712272+0000 mon.a (mon.0) 313 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:00.712272+0000 mon.a (mon.0) 313 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:00.713204+0000 mon.a (mon.0) 314 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:00.713204+0000 mon.a (mon.0) 314 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:00.713706+0000 mon.a (mon.0) 315 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:00.713706+0000 mon.a (mon.0) 315 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:00.714113+0000 mon.a (mon.0) 316 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:00.714113+0000 mon.a (mon.0) 316 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:01.072905+0000 mon.a (mon.0) 317 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:01.072905+0000 mon.a (mon.0) 317 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:01.077354+0000 mon.a (mon.0) 318 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:01.077354+0000 mon.a (mon.0) 318 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:01.078602+0000 mon.a (mon.0) 319 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:01.078602+0000 mon.a (mon.0) 319 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:01.078984+0000 mon.a (mon.0) 320 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:01.078984+0000 mon.a (mon.0) 320 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:01.079347+0000 mon.a (mon.0) 321 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:01.079347+0000 mon.a (mon.0) 321 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:01.425733+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:01.425733+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:01.429484+0000 mon.a (mon.0) 323 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:01 vm09 bash[20916]: audit 2026-03-10T10:16:01.429484+0000 mon.a (mon.0) 323 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cluster 2026-03-10T10:15:59.766573+0000 mgr.a (mgr.14150) 67 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cluster 2026-03-10T10:15:59.766573+0000 mgr.a (mgr.14150) 67 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.218832+0000 mgr.a (mgr.14150) 68 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.218832+0000 mgr.a (mgr.14150) 68 : cephadm [INF] Updating vm00:/etc/ceph/ceph.conf 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.218921+0000 mgr.a (mgr.14150) 69 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.218921+0000 mgr.a (mgr.14150) 69 : cephadm [INF] Updating vm03:/etc/ceph/ceph.conf 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.218970+0000 mgr.a (mgr.14150) 70 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.218970+0000 mgr.a (mgr.14150) 70 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.263858+0000 mgr.a (mgr.14150) 71 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.263858+0000 mgr.a (mgr.14150) 71 : cephadm [INF] Updating vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.264184+0000 mgr.a (mgr.14150) 72 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.264184+0000 mgr.a (mgr.14150) 72 : cephadm [INF] Updating vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.264288+0000 mgr.a (mgr.14150) 73 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.264288+0000 mgr.a (mgr.14150) 73 : cephadm [INF] Updating vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/config/ceph.conf 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.339843+0000 mgr.a (mgr.14150) 74 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.339843+0000 mgr.a (mgr.14150) 74 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.341422+0000 mgr.a (mgr.14150) 75 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.341422+0000 mgr.a (mgr.14150) 75 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:00.708154+0000 mon.a (mon.0) 312 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:00.708154+0000 mon.a (mon.0) 312 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:00.712272+0000 mon.a (mon.0) 313 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:00.712272+0000 mon.a (mon.0) 313 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:00.713204+0000 mon.a (mon.0) 314 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:00.713204+0000 mon.a (mon.0) 314 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:00.713706+0000 mon.a (mon.0) 315 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:00.713706+0000 mon.a (mon.0) 315 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:00.714113+0000 mon.a (mon.0) 316 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:00.714113+0000 mon.a (mon.0) 316 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:01.072905+0000 mon.a (mon.0) 317 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:01.072905+0000 mon.a (mon.0) 317 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:01.077354+0000 mon.a (mon.0) 318 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:01.077354+0000 mon.a (mon.0) 318 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:01.078602+0000 mon.a (mon.0) 319 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:01.078602+0000 mon.a (mon.0) 319 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:01.078984+0000 mon.a (mon.0) 320 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:01.078984+0000 mon.a (mon.0) 320 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:01.079347+0000 mon.a (mon.0) 321 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:01.079347+0000 mon.a (mon.0) 321 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:01.425733+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:01.425733+0000 mon.a (mon.0) 322 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:01.429484+0000 mon.a (mon.0) 323 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:01 vm03 bash[21274]: audit 2026-03-10T10:16:01.429484+0000 mon.a (mon.0) 323 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:02.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:02 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.712796+0000 mgr.a (mgr.14150) 76 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:16:02.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:02 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.712796+0000 mgr.a (mgr.14150) 76 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:16:02.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:02 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.714736+0000 mgr.a (mgr.14150) 77 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:16:02.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:02 vm00 bash[20709]: cephadm 2026-03-10T10:16:00.714736+0000 mgr.a (mgr.14150) 77 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:16:02.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:02 vm00 bash[20709]: cephadm 2026-03-10T10:16:01.078458+0000 mgr.a (mgr.14150) 78 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-10T10:16:02.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:02 vm00 bash[20709]: cephadm 2026-03-10T10:16:01.078458+0000 mgr.a (mgr.14150) 78 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-10T10:16:02.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:02 vm00 bash[20709]: cephadm 2026-03-10T10:16:01.079819+0000 mgr.a (mgr.14150) 79 : cephadm [INF] Reconfiguring daemon mon.c on vm09 2026-03-10T10:16:02.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:02 vm00 bash[20709]: cephadm 2026-03-10T10:16:01.079819+0000 mgr.a (mgr.14150) 79 : cephadm [INF] Reconfiguring daemon mon.c on vm09 2026-03-10T10:16:03.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:02 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.712796+0000 mgr.a (mgr.14150) 76 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:16:03.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:02 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.712796+0000 mgr.a (mgr.14150) 76 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:16:03.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:02 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.714736+0000 mgr.a (mgr.14150) 77 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:16:03.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:02 vm09 bash[20916]: cephadm 2026-03-10T10:16:00.714736+0000 mgr.a (mgr.14150) 77 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:16:03.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:02 vm09 bash[20916]: cephadm 2026-03-10T10:16:01.078458+0000 mgr.a (mgr.14150) 78 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-10T10:16:03.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:02 vm09 bash[20916]: cephadm 2026-03-10T10:16:01.078458+0000 mgr.a (mgr.14150) 78 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-10T10:16:03.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:02 vm09 bash[20916]: cephadm 2026-03-10T10:16:01.079819+0000 mgr.a (mgr.14150) 79 : cephadm [INF] Reconfiguring daemon mon.c on vm09 2026-03-10T10:16:03.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:02 vm09 bash[20916]: cephadm 2026-03-10T10:16:01.079819+0000 mgr.a (mgr.14150) 79 : cephadm [INF] Reconfiguring daemon mon.c on vm09 2026-03-10T10:16:03.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:02 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.712796+0000 mgr.a (mgr.14150) 76 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:16:03.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:02 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.712796+0000 mgr.a (mgr.14150) 76 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:16:03.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:02 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.714736+0000 mgr.a (mgr.14150) 77 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:16:03.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:02 vm03 bash[21274]: cephadm 2026-03-10T10:16:00.714736+0000 mgr.a (mgr.14150) 77 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:16:03.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:02 vm03 bash[21274]: cephadm 2026-03-10T10:16:01.078458+0000 mgr.a (mgr.14150) 78 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-10T10:16:03.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:02 vm03 bash[21274]: cephadm 2026-03-10T10:16:01.078458+0000 mgr.a (mgr.14150) 78 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-10T10:16:03.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:02 vm03 bash[21274]: cephadm 2026-03-10T10:16:01.079819+0000 mgr.a (mgr.14150) 79 : cephadm [INF] Reconfiguring daemon mon.c on vm09 2026-03-10T10:16:03.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:02 vm03 bash[21274]: cephadm 2026-03-10T10:16:01.079819+0000 mgr.a (mgr.14150) 79 : cephadm [INF] Reconfiguring daemon mon.c on vm09 2026-03-10T10:16:03.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:03 vm00 bash[20709]: cluster 2026-03-10T10:16:01.766939+0000 mgr.a (mgr.14150) 80 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:03.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:03 vm00 bash[20709]: cluster 2026-03-10T10:16:01.766939+0000 mgr.a (mgr.14150) 80 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:04.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:03 vm09 bash[20916]: cluster 2026-03-10T10:16:01.766939+0000 mgr.a (mgr.14150) 80 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:04.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:03 vm09 bash[20916]: cluster 2026-03-10T10:16:01.766939+0000 mgr.a (mgr.14150) 80 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:04.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:03 vm03 bash[21274]: cluster 2026-03-10T10:16:01.766939+0000 mgr.a (mgr.14150) 80 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:04.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:03 vm03 bash[21274]: cluster 2026-03-10T10:16:01.766939+0000 mgr.a (mgr.14150) 80 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:04.397 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:16:04.615 INFO:teuthology.orchestra.run.vm00.stdout:# minimal ceph.conf for 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:16:04.615 INFO:teuthology.orchestra.run.vm00.stdout:[global] 2026-03-10T10:16:04.615 INFO:teuthology.orchestra.run.vm00.stdout: fsid = 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:16:04.615 INFO:teuthology.orchestra.run.vm00.stdout: mon_host = [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] 2026-03-10T10:16:04.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:04 vm00 bash[20709]: audit 2026-03-10T10:16:04.615589+0000 mon.b (mon.2) 2 : audit [DBG] from='client.? 192.168.123.100:0/1433304707' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:04.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:04 vm00 bash[20709]: audit 2026-03-10T10:16:04.615589+0000 mon.b (mon.2) 2 : audit [DBG] from='client.? 192.168.123.100:0/1433304707' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:04.671 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-10T10:16:04.671 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-10T10:16:04.671 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T10:16:04.677 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-10T10:16:04.678 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:16:04.726 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-10T10:16:04.726 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T10:16:04.733 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-10T10:16:04.733 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:16:04.780 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T10:16:04.780 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T10:16:04.787 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T10:16:04.787 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:16:04.837 INFO:tasks.cephadm:Adding mgr.a on vm00 2026-03-10T10:16:04.837 INFO:tasks.cephadm:Adding mgr.b on vm03 2026-03-10T10:16:04.837 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph orch apply mgr '2;vm00=a;vm03=b' 2026-03-10T10:16:05.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:04 vm09 bash[20916]: audit 2026-03-10T10:16:04.615589+0000 mon.b (mon.2) 2 : audit [DBG] from='client.? 192.168.123.100:0/1433304707' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:05.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:04 vm09 bash[20916]: audit 2026-03-10T10:16:04.615589+0000 mon.b (mon.2) 2 : audit [DBG] from='client.? 192.168.123.100:0/1433304707' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:05.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:04 vm03 bash[21274]: audit 2026-03-10T10:16:04.615589+0000 mon.b (mon.2) 2 : audit [DBG] from='client.? 192.168.123.100:0/1433304707' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:05.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:04 vm03 bash[21274]: audit 2026-03-10T10:16:04.615589+0000 mon.b (mon.2) 2 : audit [DBG] from='client.? 192.168.123.100:0/1433304707' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:05.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:05 vm00 bash[20709]: cluster 2026-03-10T10:16:03.767140+0000 mgr.a (mgr.14150) 81 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:05.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:05 vm00 bash[20709]: cluster 2026-03-10T10:16:03.767140+0000 mgr.a (mgr.14150) 81 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:06.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:05 vm09 bash[20916]: cluster 2026-03-10T10:16:03.767140+0000 mgr.a (mgr.14150) 81 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:06.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:05 vm09 bash[20916]: cluster 2026-03-10T10:16:03.767140+0000 mgr.a (mgr.14150) 81 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:06.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:05 vm03 bash[21274]: cluster 2026-03-10T10:16:03.767140+0000 mgr.a (mgr.14150) 81 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:06.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:05 vm03 bash[21274]: cluster 2026-03-10T10:16:03.767140+0000 mgr.a (mgr.14150) 81 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:07.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:07 vm00 bash[20709]: cluster 2026-03-10T10:16:05.767312+0000 mgr.a (mgr.14150) 82 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:07.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:07 vm00 bash[20709]: cluster 2026-03-10T10:16:05.767312+0000 mgr.a (mgr.14150) 82 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:07.971 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:07 vm09 bash[20916]: cluster 2026-03-10T10:16:05.767312+0000 mgr.a (mgr.14150) 82 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:07.971 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:07 vm09 bash[20916]: cluster 2026-03-10T10:16:05.767312+0000 mgr.a (mgr.14150) 82 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:08.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:07 vm03 bash[21274]: cluster 2026-03-10T10:16:05.767312+0000 mgr.a (mgr.14150) 82 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:08.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:07 vm03 bash[21274]: cluster 2026-03-10T10:16:05.767312+0000 mgr.a (mgr.14150) 82 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:08.479 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.c/config 2026-03-10T10:16:08.736 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled mgr update... 2026-03-10T10:16:08.788 DEBUG:teuthology.orchestra.run.vm03:mgr.b> sudo journalctl -f -n 0 -u ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mgr.b.service 2026-03-10T10:16:08.789 INFO:tasks.cephadm:Deploying OSDs... 2026-03-10T10:16:08.789 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-10T10:16:08.789 DEBUG:teuthology.orchestra.run.vm00:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T10:16:08.792 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:16:08.792 DEBUG:teuthology.orchestra.run.vm00:> ls /dev/[sv]d? 2026-03-10T10:16:08.838 INFO:teuthology.orchestra.run.vm00.stdout:/dev/vda 2026-03-10T10:16:08.838 INFO:teuthology.orchestra.run.vm00.stdout:/dev/vdb 2026-03-10T10:16:08.838 INFO:teuthology.orchestra.run.vm00.stdout:/dev/vdc 2026-03-10T10:16:08.838 INFO:teuthology.orchestra.run.vm00.stdout:/dev/vdd 2026-03-10T10:16:08.838 INFO:teuthology.orchestra.run.vm00.stdout:/dev/vde 2026-03-10T10:16:08.838 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T10:16:08.838 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T10:16:08.838 DEBUG:teuthology.orchestra.run.vm00:> stat /dev/vdb 2026-03-10T10:16:08.882 INFO:teuthology.orchestra.run.vm00.stdout: File: /dev/vdb 2026-03-10T10:16:08.882 INFO:teuthology.orchestra.run.vm00.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T10:16:08.882 INFO:teuthology.orchestra.run.vm00.stdout:Device: 5h/5d Inode: 24 Links: 1 Device type: fe,10 2026-03-10T10:16:08.882 INFO:teuthology.orchestra.run.vm00.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:16:08.882 INFO:teuthology.orchestra.run.vm00.stdout:Access: 2026-03-10 10:07:10.398034292 +0000 2026-03-10T10:16:08.882 INFO:teuthology.orchestra.run.vm00.stdout:Modify: 2026-03-10 10:07:09.542034292 +0000 2026-03-10T10:16:08.882 INFO:teuthology.orchestra.run.vm00.stdout:Change: 2026-03-10 10:07:09.542034292 +0000 2026-03-10T10:16:08.882 INFO:teuthology.orchestra.run.vm00.stdout: Birth: - 2026-03-10T10:16:08.882 DEBUG:teuthology.orchestra.run.vm00:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T10:16:08.929 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records in 2026-03-10T10:16:08.929 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records out 2026-03-10T10:16:08.929 INFO:teuthology.orchestra.run.vm00.stderr:512 bytes copied, 0.000143228 s, 3.6 MB/s 2026-03-10T10:16:08.930 DEBUG:teuthology.orchestra.run.vm00:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T10:16:08.975 DEBUG:teuthology.orchestra.run.vm00:> stat /dev/vdc 2026-03-10T10:16:09.022 INFO:teuthology.orchestra.run.vm00.stdout: File: /dev/vdc 2026-03-10T10:16:09.022 INFO:teuthology.orchestra.run.vm00.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T10:16:09.022 INFO:teuthology.orchestra.run.vm00.stdout:Device: 5h/5d Inode: 25 Links: 1 Device type: fe,20 2026-03-10T10:16:09.022 INFO:teuthology.orchestra.run.vm00.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:16:09.023 INFO:teuthology.orchestra.run.vm00.stdout:Access: 2026-03-10 10:07:10.406034292 +0000 2026-03-10T10:16:09.023 INFO:teuthology.orchestra.run.vm00.stdout:Modify: 2026-03-10 10:07:09.530034292 +0000 2026-03-10T10:16:09.023 INFO:teuthology.orchestra.run.vm00.stdout:Change: 2026-03-10 10:07:09.530034292 +0000 2026-03-10T10:16:09.023 INFO:teuthology.orchestra.run.vm00.stdout: Birth: - 2026-03-10T10:16:09.023 DEBUG:teuthology.orchestra.run.vm00:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T10:16:09.069 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records in 2026-03-10T10:16:09.069 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records out 2026-03-10T10:16:09.069 INFO:teuthology.orchestra.run.vm00.stderr:512 bytes copied, 0.000125395 s, 4.1 MB/s 2026-03-10T10:16:09.070 DEBUG:teuthology.orchestra.run.vm00:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T10:16:09.115 DEBUG:teuthology.orchestra.run.vm00:> stat /dev/vdd 2026-03-10T10:16:09.162 INFO:teuthology.orchestra.run.vm00.stdout: File: /dev/vdd 2026-03-10T10:16:09.162 INFO:teuthology.orchestra.run.vm00.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T10:16:09.162 INFO:teuthology.orchestra.run.vm00.stdout:Device: 5h/5d Inode: 26 Links: 1 Device type: fe,30 2026-03-10T10:16:09.162 INFO:teuthology.orchestra.run.vm00.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:16:09.162 INFO:teuthology.orchestra.run.vm00.stdout:Access: 2026-03-10 10:07:10.398034292 +0000 2026-03-10T10:16:09.162 INFO:teuthology.orchestra.run.vm00.stdout:Modify: 2026-03-10 10:07:09.530034292 +0000 2026-03-10T10:16:09.162 INFO:teuthology.orchestra.run.vm00.stdout:Change: 2026-03-10 10:07:09.530034292 +0000 2026-03-10T10:16:09.162 INFO:teuthology.orchestra.run.vm00.stdout: Birth: - 2026-03-10T10:16:09.162 DEBUG:teuthology.orchestra.run.vm00:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T10:16:09.210 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records in 2026-03-10T10:16:09.210 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records out 2026-03-10T10:16:09.210 INFO:teuthology.orchestra.run.vm00.stderr:512 bytes copied, 0.000174967 s, 2.9 MB/s 2026-03-10T10:16:09.211 DEBUG:teuthology.orchestra.run.vm00:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T10:16:09.255 DEBUG:teuthology.orchestra.run.vm00:> stat /dev/vde 2026-03-10T10:16:09.280 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:16:09.280 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:09 vm03 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:16:09.302 INFO:teuthology.orchestra.run.vm00.stdout: File: /dev/vde 2026-03-10T10:16:09.302 INFO:teuthology.orchestra.run.vm00.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T10:16:09.302 INFO:teuthology.orchestra.run.vm00.stdout:Device: 5h/5d Inode: 27 Links: 1 Device type: fe,40 2026-03-10T10:16:09.302 INFO:teuthology.orchestra.run.vm00.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:16:09.302 INFO:teuthology.orchestra.run.vm00.stdout:Access: 2026-03-10 10:07:10.406034292 +0000 2026-03-10T10:16:09.302 INFO:teuthology.orchestra.run.vm00.stdout:Modify: 2026-03-10 10:07:09.530034292 +0000 2026-03-10T10:16:09.302 INFO:teuthology.orchestra.run.vm00.stdout:Change: 2026-03-10 10:07:09.530034292 +0000 2026-03-10T10:16:09.302 INFO:teuthology.orchestra.run.vm00.stdout: Birth: - 2026-03-10T10:16:09.302 DEBUG:teuthology.orchestra.run.vm00:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T10:16:09.349 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records in 2026-03-10T10:16:09.349 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records out 2026-03-10T10:16:09.349 INFO:teuthology.orchestra.run.vm00.stderr:512 bytes copied, 0.000162524 s, 3.2 MB/s 2026-03-10T10:16:09.349 DEBUG:teuthology.orchestra.run.vm00:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T10:16:09.394 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-10T10:16:09.394 DEBUG:teuthology.orchestra.run.vm03:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T10:16:09.397 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:16:09.397 DEBUG:teuthology.orchestra.run.vm03:> ls /dev/[sv]d? 2026-03-10T10:16:09.440 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vda 2026-03-10T10:16:09.440 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vdb 2026-03-10T10:16:09.440 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vdc 2026-03-10T10:16:09.440 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vdd 2026-03-10T10:16:09.440 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vde 2026-03-10T10:16:09.441 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T10:16:09.441 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T10:16:09.441 DEBUG:teuthology.orchestra.run.vm03:> stat /dev/vdb 2026-03-10T10:16:09.495 INFO:teuthology.orchestra.run.vm03.stdout: File: /dev/vdb 2026-03-10T10:16:09.495 INFO:teuthology.orchestra.run.vm03.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T10:16:09.495 INFO:teuthology.orchestra.run.vm03.stdout:Device: 5h/5d Inode: 24 Links: 1 Device type: fe,10 2026-03-10T10:16:09.495 INFO:teuthology.orchestra.run.vm03.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:16:09.495 INFO:teuthology.orchestra.run.vm03.stdout:Access: 2026-03-10 10:07:35.575263777 +0000 2026-03-10T10:16:09.495 INFO:teuthology.orchestra.run.vm03.stdout:Modify: 2026-03-10 10:07:34.579263777 +0000 2026-03-10T10:16:09.495 INFO:teuthology.orchestra.run.vm03.stdout:Change: 2026-03-10 10:07:34.579263777 +0000 2026-03-10T10:16:09.495 INFO:teuthology.orchestra.run.vm03.stdout: Birth: - 2026-03-10T10:16:09.495 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T10:16:09.544 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:16:09.546 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:09 vm03 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:16:09.546 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:09 vm03 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:16:09.546 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:09 vm03 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:16:09.546 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:09 vm03 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:16:09.546 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:09 vm03 systemd[1]: Started Ceph mgr.b for 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad. 2026-03-10T10:16:09.546 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records in 2026-03-10T10:16:09.546 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records out 2026-03-10T10:16:09.546 INFO:teuthology.orchestra.run.vm03.stderr:512 bytes copied, 0.000710289 s, 721 kB/s 2026-03-10T10:16:09.546 DEBUG:teuthology.orchestra.run.vm03:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T10:16:09.595 DEBUG:teuthology.orchestra.run.vm03:> stat /dev/vdc 2026-03-10T10:16:09.640 INFO:teuthology.orchestra.run.vm03.stdout: File: /dev/vdc 2026-03-10T10:16:09.640 INFO:teuthology.orchestra.run.vm03.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T10:16:09.640 INFO:teuthology.orchestra.run.vm03.stdout:Device: 5h/5d Inode: 25 Links: 1 Device type: fe,20 2026-03-10T10:16:09.640 INFO:teuthology.orchestra.run.vm03.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:16:09.640 INFO:teuthology.orchestra.run.vm03.stdout:Access: 2026-03-10 10:07:35.587263777 +0000 2026-03-10T10:16:09.640 INFO:teuthology.orchestra.run.vm03.stdout:Modify: 2026-03-10 10:07:34.547263777 +0000 2026-03-10T10:16:09.640 INFO:teuthology.orchestra.run.vm03.stdout:Change: 2026-03-10 10:07:34.547263777 +0000 2026-03-10T10:16:09.640 INFO:teuthology.orchestra.run.vm03.stdout: Birth: - 2026-03-10T10:16:09.641 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: cluster 2026-03-10T10:16:07.767531+0000 mgr.a (mgr.14150) 83 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: cluster 2026-03-10T10:16:07.767531+0000 mgr.a (mgr.14150) 83 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.733728+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.733728+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.734567+0000 mon.a (mon.0) 325 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.734567+0000 mon.a (mon.0) 325 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.736441+0000 mon.a (mon.0) 326 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.736441+0000 mon.a (mon.0) 326 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.737312+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.737312+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.742095+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.742095+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.743380+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.743380+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.746814+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.746814+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.748947+0000 mon.a (mon.0) 331 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.748947+0000 mon.a (mon.0) 331 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.750126+0000 mon.a (mon.0) 332 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:08.750126+0000 mon.a (mon.0) 332 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.502547+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.502547+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.508377+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.508377+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.512779+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.512779+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.516804+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.516804+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.530293+0000 mon.a (mon.0) 337 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.530293+0000 mon.a (mon.0) 337 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.531637+0000 mon.a (mon.0) 338 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.531637+0000 mon.a (mon.0) 338 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.532248+0000 mon.a (mon.0) 339 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.532248+0000 mon.a (mon.0) 339 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.536983+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.662 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.536983+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.663 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.552129+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:16:09.663 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.552129+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:16:09.663 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.552790+0000 mon.a (mon.0) 342 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:16:09.663 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.552790+0000 mon.a (mon.0) 342 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:16:09.663 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.553212+0000 mon.a (mon.0) 343 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:09.663 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.553212+0000 mon.a (mon.0) 343 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:09.663 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.571479+0000 mon.a (mon.0) 344 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.663 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:09 vm00 bash[20709]: audit 2026-03-10T10:16:09.571479+0000 mon.a (mon.0) 344 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.688 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records in 2026-03-10T10:16:09.688 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records out 2026-03-10T10:16:09.688 INFO:teuthology.orchestra.run.vm03.stderr:512 bytes copied, 0.000123771 s, 4.1 MB/s 2026-03-10T10:16:09.688 DEBUG:teuthology.orchestra.run.vm03:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T10:16:09.735 DEBUG:teuthology.orchestra.run.vm03:> stat /dev/vdd 2026-03-10T10:16:09.780 INFO:teuthology.orchestra.run.vm03.stdout: File: /dev/vdd 2026-03-10T10:16:09.780 INFO:teuthology.orchestra.run.vm03.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T10:16:09.780 INFO:teuthology.orchestra.run.vm03.stdout:Device: 5h/5d Inode: 26 Links: 1 Device type: fe,30 2026-03-10T10:16:09.781 INFO:teuthology.orchestra.run.vm03.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:16:09.781 INFO:teuthology.orchestra.run.vm03.stdout:Access: 2026-03-10 10:07:35.575263777 +0000 2026-03-10T10:16:09.781 INFO:teuthology.orchestra.run.vm03.stdout:Modify: 2026-03-10 10:07:34.575263777 +0000 2026-03-10T10:16:09.781 INFO:teuthology.orchestra.run.vm03.stdout:Change: 2026-03-10 10:07:34.575263777 +0000 2026-03-10T10:16:09.781 INFO:teuthology.orchestra.run.vm03.stdout: Birth: - 2026-03-10T10:16:09.781 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: cluster 2026-03-10T10:16:07.767531+0000 mgr.a (mgr.14150) 83 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: cluster 2026-03-10T10:16:07.767531+0000 mgr.a (mgr.14150) 83 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.733728+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.733728+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.734567+0000 mon.a (mon.0) 325 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.734567+0000 mon.a (mon.0) 325 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.736441+0000 mon.a (mon.0) 326 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.736441+0000 mon.a (mon.0) 326 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.737312+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.737312+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.742095+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.742095+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.743380+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.743380+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.746814+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.746814+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.748947+0000 mon.a (mon.0) 331 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.748947+0000 mon.a (mon.0) 331 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.750126+0000 mon.a (mon.0) 332 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:08.750126+0000 mon.a (mon.0) 332 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.502547+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.502547+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.508377+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.508377+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.512779+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.512779+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.516804+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.829 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.516804+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.530293+0000 mon.a (mon.0) 337 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.530293+0000 mon.a (mon.0) 337 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.531637+0000 mon.a (mon.0) 338 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.531637+0000 mon.a (mon.0) 338 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.532248+0000 mon.a (mon.0) 339 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.532248+0000 mon.a (mon.0) 339 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.536983+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.536983+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.552129+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.552129+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.552790+0000 mon.a (mon.0) 342 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.552790+0000 mon.a (mon.0) 342 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.553212+0000 mon.a (mon.0) 343 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.553212+0000 mon.a (mon.0) 343 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.571479+0000 mon.a (mon.0) 344 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[21274]: audit 2026-03-10T10:16:09.571479+0000 mon.a (mon.0) 344 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[22129]: debug 2026-03-10T10:16:09.618+0000 7fa1622f0640 1 -- 192.168.123.103:0/1456816679 <== mon.2 v2:192.168.123.103:3300/0 4 ==== auth_reply(proto 2 0 (0) Success) ==== 194+0+0 (secure 0 0 0) 0x55a2632614a0 con 0x55a263262800 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[22129]: debug 2026-03-10T10:16:09.678+0000 7fa164b5b140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T10:16:09.830 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[22129]: debug 2026-03-10T10:16:09.714+0000 7fa164b5b140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T10:16:09.830 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records in 2026-03-10T10:16:09.831 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records out 2026-03-10T10:16:09.831 INFO:teuthology.orchestra.run.vm03.stderr:512 bytes copied, 0.000435305 s, 1.2 MB/s 2026-03-10T10:16:09.831 DEBUG:teuthology.orchestra.run.vm03:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T10:16:09.879 DEBUG:teuthology.orchestra.run.vm03:> stat /dev/vde 2026-03-10T10:16:09.925 INFO:teuthology.orchestra.run.vm03.stdout: File: /dev/vde 2026-03-10T10:16:09.925 INFO:teuthology.orchestra.run.vm03.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T10:16:09.925 INFO:teuthology.orchestra.run.vm03.stdout:Device: 5h/5d Inode: 27 Links: 1 Device type: fe,40 2026-03-10T10:16:09.925 INFO:teuthology.orchestra.run.vm03.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:16:09.925 INFO:teuthology.orchestra.run.vm03.stdout:Access: 2026-03-10 10:07:35.587263777 +0000 2026-03-10T10:16:09.925 INFO:teuthology.orchestra.run.vm03.stdout:Modify: 2026-03-10 10:07:34.591263777 +0000 2026-03-10T10:16:09.925 INFO:teuthology.orchestra.run.vm03.stdout:Change: 2026-03-10 10:07:34.591263777 +0000 2026-03-10T10:16:09.925 INFO:teuthology.orchestra.run.vm03.stdout: Birth: - 2026-03-10T10:16:09.925 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T10:16:09.972 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records in 2026-03-10T10:16:09.973 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records out 2026-03-10T10:16:09.973 INFO:teuthology.orchestra.run.vm03.stderr:512 bytes copied, 0.000161151 s, 3.2 MB/s 2026-03-10T10:16:09.973 DEBUG:teuthology.orchestra.run.vm03:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T10:16:10.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: cluster 2026-03-10T10:16:07.767531+0000 mgr.a (mgr.14150) 83 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:10.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: cluster 2026-03-10T10:16:07.767531+0000 mgr.a (mgr.14150) 83 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:10.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.733728+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.733728+0000 mon.a (mon.0) 324 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.734567+0000 mon.a (mon.0) 325 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:10.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.734567+0000 mon.a (mon.0) 325 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:10.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.736441+0000 mon.a (mon.0) 326 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:10.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.736441+0000 mon.a (mon.0) 326 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:10.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.737312+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:10.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.737312+0000 mon.a (mon.0) 327 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:10.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.742095+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.742095+0000 mon.a (mon.0) 328 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.743380+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.743380+0000 mon.a (mon.0) 329 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.746814+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.746814+0000 mon.a (mon.0) 330 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.748947+0000 mon.a (mon.0) 331 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.748947+0000 mon.a (mon.0) 331 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.750126+0000 mon.a (mon.0) 332 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:08.750126+0000 mon.a (mon.0) 332 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.502547+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.502547+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.508377+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.508377+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.512779+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.512779+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.516804+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.516804+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.530293+0000 mon.a (mon.0) 337 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.530293+0000 mon.a (mon.0) 337 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.531637+0000 mon.a (mon.0) 338 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.531637+0000 mon.a (mon.0) 338 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.532248+0000 mon.a (mon.0) 339 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.532248+0000 mon.a (mon.0) 339 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.536983+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.536983+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.552129+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.552129+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.552790+0000 mon.a (mon.0) 342 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.552790+0000 mon.a (mon.0) 342 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.553212+0000 mon.a (mon.0) 343 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.553212+0000 mon.a (mon.0) 343 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.571479+0000 mon.a (mon.0) 344 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:09 vm09 bash[20916]: audit 2026-03-10T10:16:09.571479+0000 mon.a (mon.0) 344 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.020 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T10:16:10.020 DEBUG:teuthology.orchestra.run.vm09:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T10:16:10.022 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:16:10.023 DEBUG:teuthology.orchestra.run.vm09:> ls /dev/[sv]d? 2026-03-10T10:16:10.069 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vda 2026-03-10T10:16:10.069 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdb 2026-03-10T10:16:10.069 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdc 2026-03-10T10:16:10.069 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdd 2026-03-10T10:16:10.069 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vde 2026-03-10T10:16:10.070 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T10:16:10.070 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T10:16:10.070 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdb 2026-03-10T10:16:10.080 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:09 vm03 bash[22129]: debug 2026-03-10T10:16:09.830+0000 7fa164b5b140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T10:16:10.113 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdb 2026-03-10T10:16:10.113 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T10:16:10.113 INFO:teuthology.orchestra.run.vm09.stdout:Device: 5h/5d Inode: 24 Links: 1 Device type: fe,10 2026-03-10T10:16:10.113 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:16:10.113 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 10:08:01.720346641 +0000 2026-03-10T10:16:10.113 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 10:08:00.296346641 +0000 2026-03-10T10:16:10.113 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 10:08:00.296346641 +0000 2026-03-10T10:16:10.113 INFO:teuthology.orchestra.run.vm09.stdout: Birth: - 2026-03-10T10:16:10.113 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T10:16:10.117 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[22129]: debug 2026-03-10T10:16:10.114+0000 7fa164b5b140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T10:16:10.161 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T10:16:10.161 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T10:16:10.161 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000254496 s, 2.0 MB/s 2026-03-10T10:16:10.162 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T10:16:10.206 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdc 2026-03-10T10:16:10.248 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdc 2026-03-10T10:16:10.249 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T10:16:10.249 INFO:teuthology.orchestra.run.vm09.stdout:Device: 5h/5d Inode: 25 Links: 1 Device type: fe,20 2026-03-10T10:16:10.249 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:16:10.249 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 10:08:01.728346641 +0000 2026-03-10T10:16:10.249 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 10:08:00.296346641 +0000 2026-03-10T10:16:10.249 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 10:08:00.296346641 +0000 2026-03-10T10:16:10.249 INFO:teuthology.orchestra.run.vm09.stdout: Birth: - 2026-03-10T10:16:10.249 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T10:16:10.296 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T10:16:10.296 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T10:16:10.296 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000188884 s, 2.7 MB/s 2026-03-10T10:16:10.297 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T10:16:10.341 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdd 2026-03-10T10:16:10.384 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdd 2026-03-10T10:16:10.384 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T10:16:10.384 INFO:teuthology.orchestra.run.vm09.stdout:Device: 5h/5d Inode: 26 Links: 1 Device type: fe,30 2026-03-10T10:16:10.384 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:16:10.384 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 10:08:01.720346641 +0000 2026-03-10T10:16:10.384 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 10:08:00.296346641 +0000 2026-03-10T10:16:10.384 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 10:08:00.296346641 +0000 2026-03-10T10:16:10.384 INFO:teuthology.orchestra.run.vm09.stdout: Birth: - 2026-03-10T10:16:10.385 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T10:16:10.432 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T10:16:10.432 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T10:16:10.432 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000205153 s, 2.5 MB/s 2026-03-10T10:16:10.433 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T10:16:10.478 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vde 2026-03-10T10:16:10.521 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vde 2026-03-10T10:16:10.521 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-10T10:16:10.521 INFO:teuthology.orchestra.run.vm09.stdout:Device: 5h/5d Inode: 27 Links: 1 Device type: fe,40 2026-03-10T10:16:10.521 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:16:10.521 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 10:08:01.728346641 +0000 2026-03-10T10:16:10.521 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 10:08:00.296346641 +0000 2026-03-10T10:16:10.521 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 10:08:00.296346641 +0000 2026-03-10T10:16:10.521 INFO:teuthology.orchestra.run.vm09.stdout: Birth: - 2026-03-10T10:16:10.521 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T10:16:10.568 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T10:16:10.568 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T10:16:10.568 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000157835 s, 3.2 MB/s 2026-03-10T10:16:10.569 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T10:16:10.614 INFO:tasks.cephadm:Deploying osd.0 on vm00 with /dev/vde... 2026-03-10T10:16:10.614 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- lvm zap /dev/vde 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: audit 2026-03-10T10:16:08.728912+0000 mgr.a (mgr.14150) 84 : audit [DBG] from='client.14202 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm00=a;vm03=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: audit 2026-03-10T10:16:08.728912+0000 mgr.a (mgr.14150) 84 : audit [DBG] from='client.14202 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm00=a;vm03=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: cephadm 2026-03-10T10:16:08.729703+0000 mgr.a (mgr.14150) 85 : cephadm [INF] Saving service mgr spec with placement vm00=a;vm03=b;count:2 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: cephadm 2026-03-10T10:16:08.729703+0000 mgr.a (mgr.14150) 85 : cephadm [INF] Saving service mgr spec with placement vm00=a;vm03=b;count:2 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: cephadm 2026-03-10T10:16:08.751040+0000 mgr.a (mgr.14150) 86 : cephadm [INF] Deploying daemon mgr.b on vm03 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: cephadm 2026-03-10T10:16:08.751040+0000 mgr.a (mgr.14150) 86 : cephadm [INF] Deploying daemon mgr.b on vm03 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: cephadm 2026-03-10T10:16:09.537232+0000 mgr.a (mgr.14150) 87 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: cephadm 2026-03-10T10:16:09.537232+0000 mgr.a (mgr.14150) 87 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: cephadm 2026-03-10T10:16:09.551853+0000 mgr.a (mgr.14150) 88 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: cephadm 2026-03-10T10:16:09.551853+0000 mgr.a (mgr.14150) 88 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: cephadm 2026-03-10T10:16:09.553757+0000 mgr.a (mgr.14150) 89 : cephadm [INF] Reconfiguring daemon mgr.a on vm00 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: cephadm 2026-03-10T10:16:09.553757+0000 mgr.a (mgr.14150) 89 : cephadm [INF] Reconfiguring daemon mgr.a on vm00 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: audit 2026-03-10T10:16:09.992376+0000 mon.a (mon.0) 345 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: audit 2026-03-10T10:16:09.992376+0000 mon.a (mon.0) 345 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: audit 2026-03-10T10:16:09.996841+0000 mon.a (mon.0) 346 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: audit 2026-03-10T10:16:09.996841+0000 mon.a (mon.0) 346 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: audit 2026-03-10T10:16:09.998215+0000 mon.a (mon.0) 347 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: audit 2026-03-10T10:16:09.998215+0000 mon.a (mon.0) 347 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: audit 2026-03-10T10:16:09.999354+0000 mon.a (mon.0) 348 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:10.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: audit 2026-03-10T10:16:09.999354+0000 mon.a (mon.0) 348 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:10.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: audit 2026-03-10T10:16:09.999754+0000 mon.a (mon.0) 349 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:10.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: audit 2026-03-10T10:16:09.999754+0000 mon.a (mon.0) 349 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:10.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: audit 2026-03-10T10:16:10.004880+0000 mon.a (mon.0) 350 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:10 vm09 bash[20916]: audit 2026-03-10T10:16:10.004880+0000 mon.a (mon.0) 350 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: audit 2026-03-10T10:16:08.728912+0000 mgr.a (mgr.14150) 84 : audit [DBG] from='client.14202 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm00=a;vm03=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: audit 2026-03-10T10:16:08.728912+0000 mgr.a (mgr.14150) 84 : audit [DBG] from='client.14202 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm00=a;vm03=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: cephadm 2026-03-10T10:16:08.729703+0000 mgr.a (mgr.14150) 85 : cephadm [INF] Saving service mgr spec with placement vm00=a;vm03=b;count:2 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: cephadm 2026-03-10T10:16:08.729703+0000 mgr.a (mgr.14150) 85 : cephadm [INF] Saving service mgr spec with placement vm00=a;vm03=b;count:2 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: cephadm 2026-03-10T10:16:08.751040+0000 mgr.a (mgr.14150) 86 : cephadm [INF] Deploying daemon mgr.b on vm03 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: cephadm 2026-03-10T10:16:08.751040+0000 mgr.a (mgr.14150) 86 : cephadm [INF] Deploying daemon mgr.b on vm03 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: cephadm 2026-03-10T10:16:09.537232+0000 mgr.a (mgr.14150) 87 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: cephadm 2026-03-10T10:16:09.537232+0000 mgr.a (mgr.14150) 87 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: cephadm 2026-03-10T10:16:09.551853+0000 mgr.a (mgr.14150) 88 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: cephadm 2026-03-10T10:16:09.551853+0000 mgr.a (mgr.14150) 88 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: cephadm 2026-03-10T10:16:09.553757+0000 mgr.a (mgr.14150) 89 : cephadm [INF] Reconfiguring daemon mgr.a on vm00 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: cephadm 2026-03-10T10:16:09.553757+0000 mgr.a (mgr.14150) 89 : cephadm [INF] Reconfiguring daemon mgr.a on vm00 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: audit 2026-03-10T10:16:09.992376+0000 mon.a (mon.0) 345 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: audit 2026-03-10T10:16:09.992376+0000 mon.a (mon.0) 345 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: audit 2026-03-10T10:16:09.996841+0000 mon.a (mon.0) 346 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: audit 2026-03-10T10:16:09.996841+0000 mon.a (mon.0) 346 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: audit 2026-03-10T10:16:09.998215+0000 mon.a (mon.0) 347 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: audit 2026-03-10T10:16:09.998215+0000 mon.a (mon.0) 347 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: audit 2026-03-10T10:16:09.999354+0000 mon.a (mon.0) 348 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: audit 2026-03-10T10:16:09.999354+0000 mon.a (mon.0) 348 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: audit 2026-03-10T10:16:09.999754+0000 mon.a (mon.0) 349 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: audit 2026-03-10T10:16:09.999754+0000 mon.a (mon.0) 349 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: audit 2026-03-10T10:16:10.004880+0000 mon.a (mon.0) 350 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[21274]: audit 2026-03-10T10:16:10.004880+0000 mon.a (mon.0) 350 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[22129]: debug 2026-03-10T10:16:10.526+0000 7fa164b5b140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[22129]: debug 2026-03-10T10:16:10.602+0000 7fa164b5b140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[22129]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[22129]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[22129]: from numpy import show_config as show_numpy_config 2026-03-10T10:16:10.831 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[22129]: debug 2026-03-10T10:16:10.722+0000 7fa164b5b140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: audit 2026-03-10T10:16:08.728912+0000 mgr.a (mgr.14150) 84 : audit [DBG] from='client.14202 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm00=a;vm03=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: audit 2026-03-10T10:16:08.728912+0000 mgr.a (mgr.14150) 84 : audit [DBG] from='client.14202 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm00=a;vm03=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: cephadm 2026-03-10T10:16:08.729703+0000 mgr.a (mgr.14150) 85 : cephadm [INF] Saving service mgr spec with placement vm00=a;vm03=b;count:2 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: cephadm 2026-03-10T10:16:08.729703+0000 mgr.a (mgr.14150) 85 : cephadm [INF] Saving service mgr spec with placement vm00=a;vm03=b;count:2 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: cephadm 2026-03-10T10:16:08.751040+0000 mgr.a (mgr.14150) 86 : cephadm [INF] Deploying daemon mgr.b on vm03 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: cephadm 2026-03-10T10:16:08.751040+0000 mgr.a (mgr.14150) 86 : cephadm [INF] Deploying daemon mgr.b on vm03 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: cephadm 2026-03-10T10:16:09.537232+0000 mgr.a (mgr.14150) 87 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: cephadm 2026-03-10T10:16:09.537232+0000 mgr.a (mgr.14150) 87 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: cephadm 2026-03-10T10:16:09.551853+0000 mgr.a (mgr.14150) 88 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: cephadm 2026-03-10T10:16:09.551853+0000 mgr.a (mgr.14150) 88 : cephadm [INF] Reconfiguring mgr.a (unknown last config time)... 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: cephadm 2026-03-10T10:16:09.553757+0000 mgr.a (mgr.14150) 89 : cephadm [INF] Reconfiguring daemon mgr.a on vm00 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: cephadm 2026-03-10T10:16:09.553757+0000 mgr.a (mgr.14150) 89 : cephadm [INF] Reconfiguring daemon mgr.a on vm00 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: audit 2026-03-10T10:16:09.992376+0000 mon.a (mon.0) 345 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: audit 2026-03-10T10:16:09.992376+0000 mon.a (mon.0) 345 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: audit 2026-03-10T10:16:09.996841+0000 mon.a (mon.0) 346 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: audit 2026-03-10T10:16:09.996841+0000 mon.a (mon.0) 346 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: audit 2026-03-10T10:16:09.998215+0000 mon.a (mon.0) 347 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: audit 2026-03-10T10:16:09.998215+0000 mon.a (mon.0) 347 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: audit 2026-03-10T10:16:09.999354+0000 mon.a (mon.0) 348 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: audit 2026-03-10T10:16:09.999354+0000 mon.a (mon.0) 348 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: audit 2026-03-10T10:16:09.999754+0000 mon.a (mon.0) 349 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: audit 2026-03-10T10:16:09.999754+0000 mon.a (mon.0) 349 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: audit 2026-03-10T10:16:10.004880+0000 mon.a (mon.0) 350 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:10.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:10 vm00 bash[20709]: audit 2026-03-10T10:16:10.004880+0000 mon.a (mon.0) 350 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:11.132 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[22129]: debug 2026-03-10T10:16:10.846+0000 7fa164b5b140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T10:16:11.132 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[22129]: debug 2026-03-10T10:16:10.882+0000 7fa164b5b140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T10:16:11.132 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[22129]: debug 2026-03-10T10:16:10.926+0000 7fa164b5b140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T10:16:11.132 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:10 vm03 bash[22129]: debug 2026-03-10T10:16:10.982+0000 7fa164b5b140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T10:16:11.132 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:11 vm03 bash[22129]: debug 2026-03-10T10:16:11.050+0000 7fa164b5b140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T10:16:11.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:11 vm03 bash[21274]: cluster 2026-03-10T10:16:09.767730+0000 mgr.a (mgr.14150) 90 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:11.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:11 vm03 bash[21274]: cluster 2026-03-10T10:16:09.767730+0000 mgr.a (mgr.14150) 90 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:11.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:11 vm03 bash[21274]: cephadm 2026-03-10T10:16:10.005302+0000 mgr.a (mgr.14150) 91 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:16:11.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:11 vm03 bash[21274]: cephadm 2026-03-10T10:16:10.005302+0000 mgr.a (mgr.14150) 91 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:16:11.830 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:11 vm03 bash[22129]: debug 2026-03-10T10:16:11.498+0000 7fa164b5b140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T10:16:11.830 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:11 vm03 bash[22129]: debug 2026-03-10T10:16:11.534+0000 7fa164b5b140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T10:16:11.830 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:11 vm03 bash[22129]: debug 2026-03-10T10:16:11.570+0000 7fa164b5b140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T10:16:12.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:11 vm09 bash[20916]: cluster 2026-03-10T10:16:09.767730+0000 mgr.a (mgr.14150) 90 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:12.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:11 vm09 bash[20916]: cluster 2026-03-10T10:16:09.767730+0000 mgr.a (mgr.14150) 90 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:12.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:11 vm09 bash[20916]: cephadm 2026-03-10T10:16:10.005302+0000 mgr.a (mgr.14150) 91 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:16:12.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:11 vm09 bash[20916]: cephadm 2026-03-10T10:16:10.005302+0000 mgr.a (mgr.14150) 91 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:16:12.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:11 vm00 bash[20709]: cluster 2026-03-10T10:16:09.767730+0000 mgr.a (mgr.14150) 90 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:12.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:11 vm00 bash[20709]: cluster 2026-03-10T10:16:09.767730+0000 mgr.a (mgr.14150) 90 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:12.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:11 vm00 bash[20709]: cephadm 2026-03-10T10:16:10.005302+0000 mgr.a (mgr.14150) 91 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:16:12.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:11 vm00 bash[20709]: cephadm 2026-03-10T10:16:10.005302+0000 mgr.a (mgr.14150) 91 : cephadm [INF] Metadata not up to date on all hosts. Skipping non agent specs 2026-03-10T10:16:12.255 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:11 vm03 bash[22129]: debug 2026-03-10T10:16:11.842+0000 7fa164b5b140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T10:16:12.255 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:11 vm03 bash[22129]: debug 2026-03-10T10:16:11.886+0000 7fa164b5b140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T10:16:12.255 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:11 vm03 bash[22129]: debug 2026-03-10T10:16:11.934+0000 7fa164b5b140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T10:16:12.255 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:12 vm03 bash[22129]: debug 2026-03-10T10:16:12.062+0000 7fa164b5b140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T10:16:12.532 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:12 vm03 bash[22129]: debug 2026-03-10T10:16:12.254+0000 7fa164b5b140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T10:16:12.532 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:12 vm03 bash[22129]: debug 2026-03-10T10:16:12.450+0000 7fa164b5b140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T10:16:12.532 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:12 vm03 bash[22129]: debug 2026-03-10T10:16:12.486+0000 7fa164b5b140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T10:16:12.830 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:12 vm03 bash[22129]: debug 2026-03-10T10:16:12.530+0000 7fa164b5b140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T10:16:12.830 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:12 vm03 bash[22129]: debug 2026-03-10T10:16:12.682+0000 7fa164b5b140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T10:16:13.330 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:16:12 vm03 bash[22129]: debug 2026-03-10T10:16:12.994+0000 7fa164b5b140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T10:16:13.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:13 vm00 bash[20709]: cluster 2026-03-10T10:16:11.767997+0000 mgr.a (mgr.14150) 92 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:13.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:13 vm00 bash[20709]: cluster 2026-03-10T10:16:11.767997+0000 mgr.a (mgr.14150) 92 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:13.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:13 vm00 bash[20709]: audit 2026-03-10T10:16:12.501240+0000 mon.a (mon.0) 351 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:13.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:13 vm00 bash[20709]: audit 2026-03-10T10:16:12.501240+0000 mon.a (mon.0) 351 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:13.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:13 vm00 bash[20709]: cluster 2026-03-10T10:16:12.999526+0000 mon.a (mon.0) 352 : cluster [DBG] Standby manager daemon b started 2026-03-10T10:16:13.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:13 vm00 bash[20709]: cluster 2026-03-10T10:16:12.999526+0000 mon.a (mon.0) 352 : cluster [DBG] Standby manager daemon b started 2026-03-10T10:16:13.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:13 vm00 bash[20709]: audit 2026-03-10T10:16:13.004584+0000 mon.c (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-10T10:16:13.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:13 vm00 bash[20709]: audit 2026-03-10T10:16:13.004584+0000 mon.c (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-10T10:16:13.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:13 vm00 bash[20709]: audit 2026-03-10T10:16:13.005118+0000 mon.c (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T10:16:13.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:13 vm00 bash[20709]: audit 2026-03-10T10:16:13.005118+0000 mon.c (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T10:16:13.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:13 vm00 bash[20709]: audit 2026-03-10T10:16:13.005803+0000 mon.c (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-10T10:16:13.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:13 vm00 bash[20709]: audit 2026-03-10T10:16:13.005803+0000 mon.c (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-10T10:16:13.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:13 vm00 bash[20709]: audit 2026-03-10T10:16:13.006290+0000 mon.c (mon.1) 6 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T10:16:13.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:13 vm00 bash[20709]: audit 2026-03-10T10:16:13.006290+0000 mon.c (mon.1) 6 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T10:16:13.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:13 vm03 bash[21274]: cluster 2026-03-10T10:16:11.767997+0000 mgr.a (mgr.14150) 92 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:13.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:13 vm03 bash[21274]: cluster 2026-03-10T10:16:11.767997+0000 mgr.a (mgr.14150) 92 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:13.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:13 vm03 bash[21274]: audit 2026-03-10T10:16:12.501240+0000 mon.a (mon.0) 351 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:13.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:13 vm03 bash[21274]: audit 2026-03-10T10:16:12.501240+0000 mon.a (mon.0) 351 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:13.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:13 vm03 bash[21274]: cluster 2026-03-10T10:16:12.999526+0000 mon.a (mon.0) 352 : cluster [DBG] Standby manager daemon b started 2026-03-10T10:16:13.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:13 vm03 bash[21274]: cluster 2026-03-10T10:16:12.999526+0000 mon.a (mon.0) 352 : cluster [DBG] Standby manager daemon b started 2026-03-10T10:16:13.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:13 vm03 bash[21274]: audit 2026-03-10T10:16:13.004584+0000 mon.c (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-10T10:16:13.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:13 vm03 bash[21274]: audit 2026-03-10T10:16:13.004584+0000 mon.c (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-10T10:16:13.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:13 vm03 bash[21274]: audit 2026-03-10T10:16:13.005118+0000 mon.c (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T10:16:13.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:13 vm03 bash[21274]: audit 2026-03-10T10:16:13.005118+0000 mon.c (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T10:16:13.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:13 vm03 bash[21274]: audit 2026-03-10T10:16:13.005803+0000 mon.c (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-10T10:16:13.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:13 vm03 bash[21274]: audit 2026-03-10T10:16:13.005803+0000 mon.c (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-10T10:16:13.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:13 vm03 bash[21274]: audit 2026-03-10T10:16:13.006290+0000 mon.c (mon.1) 6 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T10:16:13.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:13 vm03 bash[21274]: audit 2026-03-10T10:16:13.006290+0000 mon.c (mon.1) 6 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T10:16:14.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:13 vm09 bash[20916]: cluster 2026-03-10T10:16:11.767997+0000 mgr.a (mgr.14150) 92 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:14.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:13 vm09 bash[20916]: cluster 2026-03-10T10:16:11.767997+0000 mgr.a (mgr.14150) 92 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:14.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:13 vm09 bash[20916]: audit 2026-03-10T10:16:12.501240+0000 mon.a (mon.0) 351 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:14.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:13 vm09 bash[20916]: audit 2026-03-10T10:16:12.501240+0000 mon.a (mon.0) 351 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:14.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:13 vm09 bash[20916]: cluster 2026-03-10T10:16:12.999526+0000 mon.a (mon.0) 352 : cluster [DBG] Standby manager daemon b started 2026-03-10T10:16:14.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:13 vm09 bash[20916]: cluster 2026-03-10T10:16:12.999526+0000 mon.a (mon.0) 352 : cluster [DBG] Standby manager daemon b started 2026-03-10T10:16:14.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:13 vm09 bash[20916]: audit 2026-03-10T10:16:13.004584+0000 mon.c (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-10T10:16:14.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:13 vm09 bash[20916]: audit 2026-03-10T10:16:13.004584+0000 mon.c (mon.1) 3 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-10T10:16:14.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:13 vm09 bash[20916]: audit 2026-03-10T10:16:13.005118+0000 mon.c (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T10:16:14.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:13 vm09 bash[20916]: audit 2026-03-10T10:16:13.005118+0000 mon.c (mon.1) 4 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T10:16:14.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:13 vm09 bash[20916]: audit 2026-03-10T10:16:13.005803+0000 mon.c (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-10T10:16:14.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:13 vm09 bash[20916]: audit 2026-03-10T10:16:13.005803+0000 mon.c (mon.1) 5 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-10T10:16:14.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:13 vm09 bash[20916]: audit 2026-03-10T10:16:13.006290+0000 mon.c (mon.1) 6 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T10:16:14.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:13 vm09 bash[20916]: audit 2026-03-10T10:16:13.006290+0000 mon.c (mon.1) 6 : audit [DBG] from='mgr.? 192.168.123.103:0/3920566500' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T10:16:15.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:14 vm09 bash[20916]: cluster 2026-03-10T10:16:13.689294+0000 mon.a (mon.0) 353 : cluster [DBG] mgrmap e13: a(active, since 67s), standbys: b 2026-03-10T10:16:15.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:14 vm09 bash[20916]: cluster 2026-03-10T10:16:13.689294+0000 mon.a (mon.0) 353 : cluster [DBG] mgrmap e13: a(active, since 67s), standbys: b 2026-03-10T10:16:15.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:14 vm09 bash[20916]: audit 2026-03-10T10:16:13.689442+0000 mon.a (mon.0) 354 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-10T10:16:15.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:14 vm09 bash[20916]: audit 2026-03-10T10:16:13.689442+0000 mon.a (mon.0) 354 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-10T10:16:15.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:14 vm03 bash[21274]: cluster 2026-03-10T10:16:13.689294+0000 mon.a (mon.0) 353 : cluster [DBG] mgrmap e13: a(active, since 67s), standbys: b 2026-03-10T10:16:15.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:14 vm03 bash[21274]: cluster 2026-03-10T10:16:13.689294+0000 mon.a (mon.0) 353 : cluster [DBG] mgrmap e13: a(active, since 67s), standbys: b 2026-03-10T10:16:15.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:14 vm03 bash[21274]: audit 2026-03-10T10:16:13.689442+0000 mon.a (mon.0) 354 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-10T10:16:15.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:14 vm03 bash[21274]: audit 2026-03-10T10:16:13.689442+0000 mon.a (mon.0) 354 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-10T10:16:15.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:14 vm00 bash[20709]: cluster 2026-03-10T10:16:13.689294+0000 mon.a (mon.0) 353 : cluster [DBG] mgrmap e13: a(active, since 67s), standbys: b 2026-03-10T10:16:15.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:14 vm00 bash[20709]: cluster 2026-03-10T10:16:13.689294+0000 mon.a (mon.0) 353 : cluster [DBG] mgrmap e13: a(active, since 67s), standbys: b 2026-03-10T10:16:15.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:14 vm00 bash[20709]: audit 2026-03-10T10:16:13.689442+0000 mon.a (mon.0) 354 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-10T10:16:15.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:14 vm00 bash[20709]: audit 2026-03-10T10:16:13.689442+0000 mon.a (mon.0) 354 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-10T10:16:15.247 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:16:16.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:15 vm09 bash[20916]: cluster 2026-03-10T10:16:13.768216+0000 mgr.a (mgr.14150) 93 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:16.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:15 vm09 bash[20916]: cluster 2026-03-10T10:16:13.768216+0000 mgr.a (mgr.14150) 93 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:16.017 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:15 vm00 bash[20709]: cluster 2026-03-10T10:16:13.768216+0000 mgr.a (mgr.14150) 93 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:16.017 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:15 vm00 bash[20709]: cluster 2026-03-10T10:16:13.768216+0000 mgr.a (mgr.14150) 93 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:16.053 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:16:16.066 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph orch daemon add osd vm00:/dev/vde 2026-03-10T10:16:16.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:15 vm03 bash[21274]: cluster 2026-03-10T10:16:13.768216+0000 mgr.a (mgr.14150) 93 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:16.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:15 vm03 bash[21274]: cluster 2026-03-10T10:16:13.768216+0000 mgr.a (mgr.14150) 93 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:17.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:17 vm03 bash[21274]: cluster 2026-03-10T10:16:15.768417+0000 mgr.a (mgr.14150) 94 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:17.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:17 vm03 bash[21274]: cluster 2026-03-10T10:16:15.768417+0000 mgr.a (mgr.14150) 94 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:17.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:17 vm03 bash[21274]: audit 2026-03-10T10:16:17.636030+0000 mon.a (mon.0) 355 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:17.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:17 vm03 bash[21274]: audit 2026-03-10T10:16:17.636030+0000 mon.a (mon.0) 355 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:17.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:17 vm03 bash[21274]: audit 2026-03-10T10:16:17.640375+0000 mon.a (mon.0) 356 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:17.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:17 vm03 bash[21274]: audit 2026-03-10T10:16:17.640375+0000 mon.a (mon.0) 356 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:17.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:17 vm03 bash[21274]: audit 2026-03-10T10:16:17.642362+0000 mon.a (mon.0) 357 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:17.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:17 vm03 bash[21274]: audit 2026-03-10T10:16:17.642362+0000 mon.a (mon.0) 357 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:17.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:17 vm03 bash[21274]: audit 2026-03-10T10:16:17.643433+0000 mon.a (mon.0) 358 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:17.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:17 vm03 bash[21274]: audit 2026-03-10T10:16:17.643433+0000 mon.a (mon.0) 358 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:17.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:17 vm03 bash[21274]: audit 2026-03-10T10:16:17.643905+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:17.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:17 vm03 bash[21274]: audit 2026-03-10T10:16:17.643905+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:17.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:17 vm03 bash[21274]: audit 2026-03-10T10:16:17.645434+0000 mon.a (mon.0) 360 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:17.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:17 vm03 bash[21274]: audit 2026-03-10T10:16:17.645434+0000 mon.a (mon.0) 360 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:17.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:17 vm03 bash[21274]: audit 2026-03-10T10:16:17.650118+0000 mon.a (mon.0) 361 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:17.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:17 vm03 bash[21274]: audit 2026-03-10T10:16:17.650118+0000 mon.a (mon.0) 361 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:18.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:17 vm09 bash[20916]: cluster 2026-03-10T10:16:15.768417+0000 mgr.a (mgr.14150) 94 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:18.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:17 vm09 bash[20916]: cluster 2026-03-10T10:16:15.768417+0000 mgr.a (mgr.14150) 94 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:18.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:17 vm09 bash[20916]: audit 2026-03-10T10:16:17.636030+0000 mon.a (mon.0) 355 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:18.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:17 vm09 bash[20916]: audit 2026-03-10T10:16:17.636030+0000 mon.a (mon.0) 355 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:18.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:17 vm09 bash[20916]: audit 2026-03-10T10:16:17.640375+0000 mon.a (mon.0) 356 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:18.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:17 vm09 bash[20916]: audit 2026-03-10T10:16:17.640375+0000 mon.a (mon.0) 356 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:18.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:17 vm09 bash[20916]: audit 2026-03-10T10:16:17.642362+0000 mon.a (mon.0) 357 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:18.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:17 vm09 bash[20916]: audit 2026-03-10T10:16:17.642362+0000 mon.a (mon.0) 357 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:18.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:17 vm09 bash[20916]: audit 2026-03-10T10:16:17.643433+0000 mon.a (mon.0) 358 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:18.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:17 vm09 bash[20916]: audit 2026-03-10T10:16:17.643433+0000 mon.a (mon.0) 358 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:18.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:17 vm09 bash[20916]: audit 2026-03-10T10:16:17.643905+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:18.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:17 vm09 bash[20916]: audit 2026-03-10T10:16:17.643905+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:18.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:17 vm09 bash[20916]: audit 2026-03-10T10:16:17.645434+0000 mon.a (mon.0) 360 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:18.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:17 vm09 bash[20916]: audit 2026-03-10T10:16:17.645434+0000 mon.a (mon.0) 360 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:18.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:17 vm09 bash[20916]: audit 2026-03-10T10:16:17.650118+0000 mon.a (mon.0) 361 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:18.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:17 vm09 bash[20916]: audit 2026-03-10T10:16:17.650118+0000 mon.a (mon.0) 361 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:17 vm00 bash[20709]: cluster 2026-03-10T10:16:15.768417+0000 mgr.a (mgr.14150) 94 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:17 vm00 bash[20709]: cluster 2026-03-10T10:16:15.768417+0000 mgr.a (mgr.14150) 94 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:17 vm00 bash[20709]: audit 2026-03-10T10:16:17.636030+0000 mon.a (mon.0) 355 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:17 vm00 bash[20709]: audit 2026-03-10T10:16:17.636030+0000 mon.a (mon.0) 355 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:17 vm00 bash[20709]: audit 2026-03-10T10:16:17.640375+0000 mon.a (mon.0) 356 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:17 vm00 bash[20709]: audit 2026-03-10T10:16:17.640375+0000 mon.a (mon.0) 356 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:17 vm00 bash[20709]: audit 2026-03-10T10:16:17.642362+0000 mon.a (mon.0) 357 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:17 vm00 bash[20709]: audit 2026-03-10T10:16:17.642362+0000 mon.a (mon.0) 357 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:17 vm00 bash[20709]: audit 2026-03-10T10:16:17.643433+0000 mon.a (mon.0) 358 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:17 vm00 bash[20709]: audit 2026-03-10T10:16:17.643433+0000 mon.a (mon.0) 358 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:17 vm00 bash[20709]: audit 2026-03-10T10:16:17.643905+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:17 vm00 bash[20709]: audit 2026-03-10T10:16:17.643905+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:17 vm00 bash[20709]: audit 2026-03-10T10:16:17.645434+0000 mon.a (mon.0) 360 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:17 vm00 bash[20709]: audit 2026-03-10T10:16:17.645434+0000 mon.a (mon.0) 360 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:17 vm00 bash[20709]: audit 2026-03-10T10:16:17.650118+0000 mon.a (mon.0) 361 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:18.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:17 vm00 bash[20709]: audit 2026-03-10T10:16:17.650118+0000 mon.a (mon.0) 361 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: cluster 2026-03-10T10:16:17.768580+0000 mgr.a (mgr.14150) 95 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: cluster 2026-03-10T10:16:17.768580+0000 mgr.a (mgr.14150) 95 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.282636+0000 mon.a (mon.0) 362 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.282636+0000 mon.a (mon.0) 362 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.287308+0000 mon.a (mon.0) 363 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.287308+0000 mon.a (mon.0) 363 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.292024+0000 mon.a (mon.0) 364 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.292024+0000 mon.a (mon.0) 364 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.617453+0000 mon.a (mon.0) 365 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.617453+0000 mon.a (mon.0) 365 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.621826+0000 mon.a (mon.0) 366 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.621826+0000 mon.a (mon.0) 366 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.626250+0000 mon.a (mon.0) 367 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.626250+0000 mon.a (mon.0) 367 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.746855+0000 mon.a (mon.0) 368 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.746855+0000 mon.a (mon.0) 368 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.750867+0000 mon.a (mon.0) 369 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.750867+0000 mon.a (mon.0) 369 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.754995+0000 mon.a (mon.0) 370 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:19 vm03 bash[21274]: audit 2026-03-10T10:16:18.754995+0000 mon.a (mon.0) 370 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: cluster 2026-03-10T10:16:17.768580+0000 mgr.a (mgr.14150) 95 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: cluster 2026-03-10T10:16:17.768580+0000 mgr.a (mgr.14150) 95 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.282636+0000 mon.a (mon.0) 362 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.282636+0000 mon.a (mon.0) 362 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.287308+0000 mon.a (mon.0) 363 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.287308+0000 mon.a (mon.0) 363 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.292024+0000 mon.a (mon.0) 364 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.292024+0000 mon.a (mon.0) 364 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.617453+0000 mon.a (mon.0) 365 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.617453+0000 mon.a (mon.0) 365 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.621826+0000 mon.a (mon.0) 366 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.621826+0000 mon.a (mon.0) 366 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.626250+0000 mon.a (mon.0) 367 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.626250+0000 mon.a (mon.0) 367 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.746855+0000 mon.a (mon.0) 368 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.746855+0000 mon.a (mon.0) 368 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.750867+0000 mon.a (mon.0) 369 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.750867+0000 mon.a (mon.0) 369 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.754995+0000 mon.a (mon.0) 370 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:19 vm00 bash[20709]: audit 2026-03-10T10:16:18.754995+0000 mon.a (mon.0) 370 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: cluster 2026-03-10T10:16:17.768580+0000 mgr.a (mgr.14150) 95 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:19.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: cluster 2026-03-10T10:16:17.768580+0000 mgr.a (mgr.14150) 95 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:19.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.282636+0000 mon.a (mon.0) 362 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.282636+0000 mon.a (mon.0) 362 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.287308+0000 mon.a (mon.0) 363 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.287308+0000 mon.a (mon.0) 363 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.292024+0000 mon.a (mon.0) 364 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.292024+0000 mon.a (mon.0) 364 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.617453+0000 mon.a (mon.0) 365 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.617453+0000 mon.a (mon.0) 365 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.621826+0000 mon.a (mon.0) 366 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.621826+0000 mon.a (mon.0) 366 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.626250+0000 mon.a (mon.0) 367 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.626250+0000 mon.a (mon.0) 367 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.746855+0000 mon.a (mon.0) 368 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.746855+0000 mon.a (mon.0) 368 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.750867+0000 mon.a (mon.0) 369 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.750867+0000 mon.a (mon.0) 369 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.754995+0000 mon.a (mon.0) 370 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:19.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:19 vm09 bash[20916]: audit 2026-03-10T10:16:18.754995+0000 mon.a (mon.0) 370 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:20.674 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:16:21.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:21 vm03 bash[21274]: cluster 2026-03-10T10:16:19.768804+0000 mgr.a (mgr.14150) 96 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:21.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:21 vm03 bash[21274]: cluster 2026-03-10T10:16:19.768804+0000 mgr.a (mgr.14150) 96 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:21.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:21 vm03 bash[21274]: audit 2026-03-10T10:16:20.918171+0000 mon.a (mon.0) 371 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:16:21.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:21 vm03 bash[21274]: audit 2026-03-10T10:16:20.918171+0000 mon.a (mon.0) 371 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:16:21.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:21 vm03 bash[21274]: audit 2026-03-10T10:16:20.919468+0000 mon.a (mon.0) 372 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:16:21.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:21 vm03 bash[21274]: audit 2026-03-10T10:16:20.919468+0000 mon.a (mon.0) 372 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:16:21.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:21 vm03 bash[21274]: audit 2026-03-10T10:16:20.919836+0000 mon.a (mon.0) 373 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:21.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:21 vm03 bash[21274]: audit 2026-03-10T10:16:20.919836+0000 mon.a (mon.0) 373 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:21.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:21 vm00 bash[20709]: cluster 2026-03-10T10:16:19.768804+0000 mgr.a (mgr.14150) 96 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:21.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:21 vm00 bash[20709]: cluster 2026-03-10T10:16:19.768804+0000 mgr.a (mgr.14150) 96 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:21.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:21 vm00 bash[20709]: audit 2026-03-10T10:16:20.918171+0000 mon.a (mon.0) 371 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:16:21.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:21 vm00 bash[20709]: audit 2026-03-10T10:16:20.918171+0000 mon.a (mon.0) 371 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:16:21.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:21 vm00 bash[20709]: audit 2026-03-10T10:16:20.919468+0000 mon.a (mon.0) 372 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:16:21.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:21 vm00 bash[20709]: audit 2026-03-10T10:16:20.919468+0000 mon.a (mon.0) 372 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:16:21.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:21 vm00 bash[20709]: audit 2026-03-10T10:16:20.919836+0000 mon.a (mon.0) 373 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:21.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:21 vm00 bash[20709]: audit 2026-03-10T10:16:20.919836+0000 mon.a (mon.0) 373 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:21.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:21 vm09 bash[20916]: cluster 2026-03-10T10:16:19.768804+0000 mgr.a (mgr.14150) 96 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:21.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:21 vm09 bash[20916]: cluster 2026-03-10T10:16:19.768804+0000 mgr.a (mgr.14150) 96 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:21.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:21 vm09 bash[20916]: audit 2026-03-10T10:16:20.918171+0000 mon.a (mon.0) 371 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:16:21.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:21 vm09 bash[20916]: audit 2026-03-10T10:16:20.918171+0000 mon.a (mon.0) 371 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:16:21.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:21 vm09 bash[20916]: audit 2026-03-10T10:16:20.919468+0000 mon.a (mon.0) 372 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:16:21.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:21 vm09 bash[20916]: audit 2026-03-10T10:16:20.919468+0000 mon.a (mon.0) 372 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:16:21.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:21 vm09 bash[20916]: audit 2026-03-10T10:16:20.919836+0000 mon.a (mon.0) 373 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:21.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:21 vm09 bash[20916]: audit 2026-03-10T10:16:20.919836+0000 mon.a (mon.0) 373 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:22.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:22 vm03 bash[21274]: audit 2026-03-10T10:16:20.916847+0000 mgr.a (mgr.14150) 97 : audit [DBG] from='client.24118 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm00:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:22.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:22 vm03 bash[21274]: audit 2026-03-10T10:16:20.916847+0000 mgr.a (mgr.14150) 97 : audit [DBG] from='client.24118 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm00:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:22.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:22 vm00 bash[20709]: audit 2026-03-10T10:16:20.916847+0000 mgr.a (mgr.14150) 97 : audit [DBG] from='client.24118 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm00:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:22.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:22 vm00 bash[20709]: audit 2026-03-10T10:16:20.916847+0000 mgr.a (mgr.14150) 97 : audit [DBG] from='client.24118 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm00:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:22.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:22 vm09 bash[20916]: audit 2026-03-10T10:16:20.916847+0000 mgr.a (mgr.14150) 97 : audit [DBG] from='client.24118 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm00:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:22.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:22 vm09 bash[20916]: audit 2026-03-10T10:16:20.916847+0000 mgr.a (mgr.14150) 97 : audit [DBG] from='client.24118 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm00:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:23.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:23 vm03 bash[21274]: cluster 2026-03-10T10:16:21.768983+0000 mgr.a (mgr.14150) 98 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:23.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:23 vm03 bash[21274]: cluster 2026-03-10T10:16:21.768983+0000 mgr.a (mgr.14150) 98 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:23.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:23 vm00 bash[20709]: cluster 2026-03-10T10:16:21.768983+0000 mgr.a (mgr.14150) 98 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:23.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:23 vm00 bash[20709]: cluster 2026-03-10T10:16:21.768983+0000 mgr.a (mgr.14150) 98 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:23.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:23 vm09 bash[20916]: cluster 2026-03-10T10:16:21.768983+0000 mgr.a (mgr.14150) 98 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:23.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:23 vm09 bash[20916]: cluster 2026-03-10T10:16:21.768983+0000 mgr.a (mgr.14150) 98 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:25.428 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:25 vm09 bash[20916]: cluster 2026-03-10T10:16:23.769170+0000 mgr.a (mgr.14150) 99 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:25.428 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:25 vm09 bash[20916]: cluster 2026-03-10T10:16:23.769170+0000 mgr.a (mgr.14150) 99 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:25.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:25 vm00 bash[20709]: cluster 2026-03-10T10:16:23.769170+0000 mgr.a (mgr.14150) 99 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:25.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:25 vm00 bash[20709]: cluster 2026-03-10T10:16:23.769170+0000 mgr.a (mgr.14150) 99 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:25.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:25 vm03 bash[21274]: cluster 2026-03-10T10:16:23.769170+0000 mgr.a (mgr.14150) 99 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:25.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:25 vm03 bash[21274]: cluster 2026-03-10T10:16:23.769170+0000 mgr.a (mgr.14150) 99 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:27.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:27 vm00 bash[20709]: cluster 2026-03-10T10:16:25.769373+0000 mgr.a (mgr.14150) 100 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:27.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:27 vm00 bash[20709]: cluster 2026-03-10T10:16:25.769373+0000 mgr.a (mgr.14150) 100 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:27.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:27 vm00 bash[20709]: audit 2026-03-10T10:16:26.743229+0000 mon.a (mon.0) 374 : audit [INF] from='client.? 192.168.123.100:0/1278451993' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8a37c1d2-927c-46cd-a2b4-e6451084fc61"}]: dispatch 2026-03-10T10:16:27.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:27 vm00 bash[20709]: audit 2026-03-10T10:16:26.743229+0000 mon.a (mon.0) 374 : audit [INF] from='client.? 192.168.123.100:0/1278451993' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8a37c1d2-927c-46cd-a2b4-e6451084fc61"}]: dispatch 2026-03-10T10:16:27.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:27 vm00 bash[20709]: audit 2026-03-10T10:16:26.745895+0000 mon.a (mon.0) 375 : audit [INF] from='client.? 192.168.123.100:0/1278451993' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8a37c1d2-927c-46cd-a2b4-e6451084fc61"}]': finished 2026-03-10T10:16:27.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:27 vm00 bash[20709]: audit 2026-03-10T10:16:26.745895+0000 mon.a (mon.0) 375 : audit [INF] from='client.? 192.168.123.100:0/1278451993' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8a37c1d2-927c-46cd-a2b4-e6451084fc61"}]': finished 2026-03-10T10:16:27.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:27 vm00 bash[20709]: cluster 2026-03-10T10:16:26.748058+0000 mon.a (mon.0) 376 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-10T10:16:27.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:27 vm00 bash[20709]: cluster 2026-03-10T10:16:26.748058+0000 mon.a (mon.0) 376 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-10T10:16:27.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:27 vm00 bash[20709]: audit 2026-03-10T10:16:26.748239+0000 mon.a (mon.0) 377 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:27.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:27 vm00 bash[20709]: audit 2026-03-10T10:16:26.748239+0000 mon.a (mon.0) 377 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:27.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:27 vm09 bash[20916]: cluster 2026-03-10T10:16:25.769373+0000 mgr.a (mgr.14150) 100 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:27.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:27 vm09 bash[20916]: cluster 2026-03-10T10:16:25.769373+0000 mgr.a (mgr.14150) 100 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:27.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:27 vm09 bash[20916]: audit 2026-03-10T10:16:26.743229+0000 mon.a (mon.0) 374 : audit [INF] from='client.? 192.168.123.100:0/1278451993' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8a37c1d2-927c-46cd-a2b4-e6451084fc61"}]: dispatch 2026-03-10T10:16:27.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:27 vm09 bash[20916]: audit 2026-03-10T10:16:26.743229+0000 mon.a (mon.0) 374 : audit [INF] from='client.? 192.168.123.100:0/1278451993' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8a37c1d2-927c-46cd-a2b4-e6451084fc61"}]: dispatch 2026-03-10T10:16:27.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:27 vm09 bash[20916]: audit 2026-03-10T10:16:26.745895+0000 mon.a (mon.0) 375 : audit [INF] from='client.? 192.168.123.100:0/1278451993' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8a37c1d2-927c-46cd-a2b4-e6451084fc61"}]': finished 2026-03-10T10:16:27.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:27 vm09 bash[20916]: audit 2026-03-10T10:16:26.745895+0000 mon.a (mon.0) 375 : audit [INF] from='client.? 192.168.123.100:0/1278451993' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8a37c1d2-927c-46cd-a2b4-e6451084fc61"}]': finished 2026-03-10T10:16:27.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:27 vm09 bash[20916]: cluster 2026-03-10T10:16:26.748058+0000 mon.a (mon.0) 376 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-10T10:16:27.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:27 vm09 bash[20916]: cluster 2026-03-10T10:16:26.748058+0000 mon.a (mon.0) 376 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-10T10:16:27.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:27 vm09 bash[20916]: audit 2026-03-10T10:16:26.748239+0000 mon.a (mon.0) 377 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:27.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:27 vm09 bash[20916]: audit 2026-03-10T10:16:26.748239+0000 mon.a (mon.0) 377 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:27.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:27 vm03 bash[21274]: cluster 2026-03-10T10:16:25.769373+0000 mgr.a (mgr.14150) 100 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:27.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:27 vm03 bash[21274]: cluster 2026-03-10T10:16:25.769373+0000 mgr.a (mgr.14150) 100 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:27.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:27 vm03 bash[21274]: audit 2026-03-10T10:16:26.743229+0000 mon.a (mon.0) 374 : audit [INF] from='client.? 192.168.123.100:0/1278451993' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8a37c1d2-927c-46cd-a2b4-e6451084fc61"}]: dispatch 2026-03-10T10:16:27.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:27 vm03 bash[21274]: audit 2026-03-10T10:16:26.743229+0000 mon.a (mon.0) 374 : audit [INF] from='client.? 192.168.123.100:0/1278451993' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8a37c1d2-927c-46cd-a2b4-e6451084fc61"}]: dispatch 2026-03-10T10:16:27.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:27 vm03 bash[21274]: audit 2026-03-10T10:16:26.745895+0000 mon.a (mon.0) 375 : audit [INF] from='client.? 192.168.123.100:0/1278451993' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8a37c1d2-927c-46cd-a2b4-e6451084fc61"}]': finished 2026-03-10T10:16:27.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:27 vm03 bash[21274]: audit 2026-03-10T10:16:26.745895+0000 mon.a (mon.0) 375 : audit [INF] from='client.? 192.168.123.100:0/1278451993' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8a37c1d2-927c-46cd-a2b4-e6451084fc61"}]': finished 2026-03-10T10:16:27.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:27 vm03 bash[21274]: cluster 2026-03-10T10:16:26.748058+0000 mon.a (mon.0) 376 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-10T10:16:27.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:27 vm03 bash[21274]: cluster 2026-03-10T10:16:26.748058+0000 mon.a (mon.0) 376 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-10T10:16:27.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:27 vm03 bash[21274]: audit 2026-03-10T10:16:26.748239+0000 mon.a (mon.0) 377 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:27.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:27 vm03 bash[21274]: audit 2026-03-10T10:16:26.748239+0000 mon.a (mon.0) 377 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:28.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:28 vm00 bash[20709]: audit 2026-03-10T10:16:27.410649+0000 mon.a (mon.0) 378 : audit [DBG] from='client.? 192.168.123.100:0/4144461830' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:16:28.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:28 vm00 bash[20709]: audit 2026-03-10T10:16:27.410649+0000 mon.a (mon.0) 378 : audit [DBG] from='client.? 192.168.123.100:0/4144461830' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:16:28.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:28 vm09 bash[20916]: audit 2026-03-10T10:16:27.410649+0000 mon.a (mon.0) 378 : audit [DBG] from='client.? 192.168.123.100:0/4144461830' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:16:28.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:28 vm09 bash[20916]: audit 2026-03-10T10:16:27.410649+0000 mon.a (mon.0) 378 : audit [DBG] from='client.? 192.168.123.100:0/4144461830' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:16:28.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:28 vm03 bash[21274]: audit 2026-03-10T10:16:27.410649+0000 mon.a (mon.0) 378 : audit [DBG] from='client.? 192.168.123.100:0/4144461830' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:16:28.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:28 vm03 bash[21274]: audit 2026-03-10T10:16:27.410649+0000 mon.a (mon.0) 378 : audit [DBG] from='client.? 192.168.123.100:0/4144461830' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:16:29.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:29 vm00 bash[20709]: cluster 2026-03-10T10:16:27.769614+0000 mgr.a (mgr.14150) 101 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:29.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:29 vm00 bash[20709]: cluster 2026-03-10T10:16:27.769614+0000 mgr.a (mgr.14150) 101 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:29.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:29 vm09 bash[20916]: cluster 2026-03-10T10:16:27.769614+0000 mgr.a (mgr.14150) 101 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:29.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:29 vm09 bash[20916]: cluster 2026-03-10T10:16:27.769614+0000 mgr.a (mgr.14150) 101 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:29.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:29 vm03 bash[21274]: cluster 2026-03-10T10:16:27.769614+0000 mgr.a (mgr.14150) 101 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:29.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:29 vm03 bash[21274]: cluster 2026-03-10T10:16:27.769614+0000 mgr.a (mgr.14150) 101 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:31.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:31 vm03 bash[21274]: cluster 2026-03-10T10:16:29.769830+0000 mgr.a (mgr.14150) 102 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:31.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:31 vm03 bash[21274]: cluster 2026-03-10T10:16:29.769830+0000 mgr.a (mgr.14150) 102 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:31.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:31 vm00 bash[20709]: cluster 2026-03-10T10:16:29.769830+0000 mgr.a (mgr.14150) 102 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:31.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:31 vm00 bash[20709]: cluster 2026-03-10T10:16:29.769830+0000 mgr.a (mgr.14150) 102 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:32.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:31 vm09 bash[20916]: cluster 2026-03-10T10:16:29.769830+0000 mgr.a (mgr.14150) 102 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:32.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:31 vm09 bash[20916]: cluster 2026-03-10T10:16:29.769830+0000 mgr.a (mgr.14150) 102 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:33.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:33 vm00 bash[20709]: cluster 2026-03-10T10:16:31.770090+0000 mgr.a (mgr.14150) 103 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:33.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:33 vm00 bash[20709]: cluster 2026-03-10T10:16:31.770090+0000 mgr.a (mgr.14150) 103 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:33.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:33 vm03 bash[21274]: cluster 2026-03-10T10:16:31.770090+0000 mgr.a (mgr.14150) 103 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:33.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:33 vm03 bash[21274]: cluster 2026-03-10T10:16:31.770090+0000 mgr.a (mgr.14150) 103 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:34.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:33 vm09 bash[20916]: cluster 2026-03-10T10:16:31.770090+0000 mgr.a (mgr.14150) 103 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:34.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:33 vm09 bash[20916]: cluster 2026-03-10T10:16:31.770090+0000 mgr.a (mgr.14150) 103 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:35 vm00 bash[20709]: cluster 2026-03-10T10:16:33.770411+0000 mgr.a (mgr.14150) 104 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:35.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:35 vm00 bash[20709]: cluster 2026-03-10T10:16:33.770411+0000 mgr.a (mgr.14150) 104 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:35.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:35 vm03 bash[21274]: cluster 2026-03-10T10:16:33.770411+0000 mgr.a (mgr.14150) 104 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:35.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:35 vm03 bash[21274]: cluster 2026-03-10T10:16:33.770411+0000 mgr.a (mgr.14150) 104 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:36.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:35 vm09 bash[20916]: cluster 2026-03-10T10:16:33.770411+0000 mgr.a (mgr.14150) 104 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:36.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:35 vm09 bash[20916]: cluster 2026-03-10T10:16:33.770411+0000 mgr.a (mgr.14150) 104 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:36.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:36 vm00 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:16:36.531 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:36 vm00 bash[20709]: audit 2026-03-10T10:16:35.736060+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T10:16:36.531 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:36 vm00 bash[20709]: audit 2026-03-10T10:16:35.736060+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T10:16:36.531 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:36 vm00 bash[20709]: audit 2026-03-10T10:16:35.736572+0000 mon.a (mon.0) 380 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:36.531 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:36 vm00 bash[20709]: audit 2026-03-10T10:16:35.736572+0000 mon.a (mon.0) 380 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:36.531 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:36 vm00 bash[20709]: audit 2026-03-10T10:16:36.177769+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:36.531 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:36 vm00 bash[20709]: audit 2026-03-10T10:16:36.177769+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:36.531 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:36 vm00 bash[20709]: audit 2026-03-10T10:16:36.182264+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:36.531 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:36 vm00 bash[20709]: audit 2026-03-10T10:16:36.182264+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:36.531 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:36 vm00 bash[20709]: audit 2026-03-10T10:16:36.188386+0000 mon.a (mon.0) 383 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:36.531 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:36 vm00 bash[20709]: audit 2026-03-10T10:16:36.188386+0000 mon.a (mon.0) 383 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:36.531 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:16:36 vm00 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:16:36.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:36 vm03 bash[21274]: audit 2026-03-10T10:16:35.736060+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T10:16:36.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:36 vm03 bash[21274]: audit 2026-03-10T10:16:35.736060+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T10:16:36.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:36 vm03 bash[21274]: audit 2026-03-10T10:16:35.736572+0000 mon.a (mon.0) 380 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:36.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:36 vm03 bash[21274]: audit 2026-03-10T10:16:35.736572+0000 mon.a (mon.0) 380 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:36.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:36 vm03 bash[21274]: audit 2026-03-10T10:16:36.177769+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:36.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:36 vm03 bash[21274]: audit 2026-03-10T10:16:36.177769+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:36.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:36 vm03 bash[21274]: audit 2026-03-10T10:16:36.182264+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:36.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:36 vm03 bash[21274]: audit 2026-03-10T10:16:36.182264+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:36.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:36 vm03 bash[21274]: audit 2026-03-10T10:16:36.188386+0000 mon.a (mon.0) 383 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:36.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:36 vm03 bash[21274]: audit 2026-03-10T10:16:36.188386+0000 mon.a (mon.0) 383 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:36.833 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:36 vm00 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:16:36.833 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 10 10:16:36 vm00 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:16:37.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:36 vm09 bash[20916]: audit 2026-03-10T10:16:35.736060+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T10:16:37.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:36 vm09 bash[20916]: audit 2026-03-10T10:16:35.736060+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T10:16:37.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:36 vm09 bash[20916]: audit 2026-03-10T10:16:35.736572+0000 mon.a (mon.0) 380 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:37.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:36 vm09 bash[20916]: audit 2026-03-10T10:16:35.736572+0000 mon.a (mon.0) 380 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:37.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:36 vm09 bash[20916]: audit 2026-03-10T10:16:36.177769+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:37.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:36 vm09 bash[20916]: audit 2026-03-10T10:16:36.177769+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:37.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:36 vm09 bash[20916]: audit 2026-03-10T10:16:36.182264+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:37.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:36 vm09 bash[20916]: audit 2026-03-10T10:16:36.182264+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:37.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:36 vm09 bash[20916]: audit 2026-03-10T10:16:36.188386+0000 mon.a (mon.0) 383 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:37.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:36 vm09 bash[20916]: audit 2026-03-10T10:16:36.188386+0000 mon.a (mon.0) 383 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:37.789 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:37 vm00 bash[20709]: cephadm 2026-03-10T10:16:35.736946+0000 mgr.a (mgr.14150) 105 : cephadm [INF] Deploying daemon osd.0 on vm00 2026-03-10T10:16:37.789 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:37 vm00 bash[20709]: cephadm 2026-03-10T10:16:35.736946+0000 mgr.a (mgr.14150) 105 : cephadm [INF] Deploying daemon osd.0 on vm00 2026-03-10T10:16:37.789 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:37 vm00 bash[20709]: cluster 2026-03-10T10:16:35.770595+0000 mgr.a (mgr.14150) 106 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:37.789 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:37 vm00 bash[20709]: cluster 2026-03-10T10:16:35.770595+0000 mgr.a (mgr.14150) 106 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:37.789 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:37 vm00 bash[20709]: audit 2026-03-10T10:16:36.730051+0000 mon.a (mon.0) 384 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:37.789 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:37 vm00 bash[20709]: audit 2026-03-10T10:16:36.730051+0000 mon.a (mon.0) 384 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:37.789 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:37 vm00 bash[20709]: audit 2026-03-10T10:16:36.732206+0000 mon.a (mon.0) 385 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:37.789 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:37 vm00 bash[20709]: audit 2026-03-10T10:16:36.732206+0000 mon.a (mon.0) 385 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:37.789 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:37 vm00 bash[20709]: audit 2026-03-10T10:16:36.732623+0000 mon.a (mon.0) 386 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:37.789 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:37 vm00 bash[20709]: audit 2026-03-10T10:16:36.732623+0000 mon.a (mon.0) 386 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:37.789 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:37 vm00 bash[20709]: audit 2026-03-10T10:16:36.735112+0000 mon.a (mon.0) 387 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:37.789 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:37 vm00 bash[20709]: audit 2026-03-10T10:16:36.735112+0000 mon.a (mon.0) 387 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:37.789 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:37 vm00 bash[20709]: audit 2026-03-10T10:16:36.739420+0000 mon.a (mon.0) 388 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:37.789 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:37 vm00 bash[20709]: audit 2026-03-10T10:16:36.739420+0000 mon.a (mon.0) 388 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:37.789 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:37 vm00 bash[20709]: audit 2026-03-10T10:16:36.743982+0000 mon.a (mon.0) 389 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:37.789 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:37 vm00 bash[20709]: audit 2026-03-10T10:16:36.743982+0000 mon.a (mon.0) 389 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:37.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:37 vm03 bash[21274]: cephadm 2026-03-10T10:16:35.736946+0000 mgr.a (mgr.14150) 105 : cephadm [INF] Deploying daemon osd.0 on vm00 2026-03-10T10:16:37.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:37 vm03 bash[21274]: cephadm 2026-03-10T10:16:35.736946+0000 mgr.a (mgr.14150) 105 : cephadm [INF] Deploying daemon osd.0 on vm00 2026-03-10T10:16:37.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:37 vm03 bash[21274]: cluster 2026-03-10T10:16:35.770595+0000 mgr.a (mgr.14150) 106 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:37.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:37 vm03 bash[21274]: cluster 2026-03-10T10:16:35.770595+0000 mgr.a (mgr.14150) 106 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:37.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:37 vm03 bash[21274]: audit 2026-03-10T10:16:36.730051+0000 mon.a (mon.0) 384 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:37.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:37 vm03 bash[21274]: audit 2026-03-10T10:16:36.730051+0000 mon.a (mon.0) 384 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:37.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:37 vm03 bash[21274]: audit 2026-03-10T10:16:36.732206+0000 mon.a (mon.0) 385 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:37.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:37 vm03 bash[21274]: audit 2026-03-10T10:16:36.732206+0000 mon.a (mon.0) 385 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:37.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:37 vm03 bash[21274]: audit 2026-03-10T10:16:36.732623+0000 mon.a (mon.0) 386 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:37.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:37 vm03 bash[21274]: audit 2026-03-10T10:16:36.732623+0000 mon.a (mon.0) 386 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:37.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:37 vm03 bash[21274]: audit 2026-03-10T10:16:36.735112+0000 mon.a (mon.0) 387 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:37.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:37 vm03 bash[21274]: audit 2026-03-10T10:16:36.735112+0000 mon.a (mon.0) 387 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:37.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:37 vm03 bash[21274]: audit 2026-03-10T10:16:36.739420+0000 mon.a (mon.0) 388 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:37.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:37 vm03 bash[21274]: audit 2026-03-10T10:16:36.739420+0000 mon.a (mon.0) 388 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:37.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:37 vm03 bash[21274]: audit 2026-03-10T10:16:36.743982+0000 mon.a (mon.0) 389 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:37.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:37 vm03 bash[21274]: audit 2026-03-10T10:16:36.743982+0000 mon.a (mon.0) 389 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:38.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:37 vm09 bash[20916]: cephadm 2026-03-10T10:16:35.736946+0000 mgr.a (mgr.14150) 105 : cephadm [INF] Deploying daemon osd.0 on vm00 2026-03-10T10:16:38.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:37 vm09 bash[20916]: cephadm 2026-03-10T10:16:35.736946+0000 mgr.a (mgr.14150) 105 : cephadm [INF] Deploying daemon osd.0 on vm00 2026-03-10T10:16:38.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:37 vm09 bash[20916]: cluster 2026-03-10T10:16:35.770595+0000 mgr.a (mgr.14150) 106 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:38.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:37 vm09 bash[20916]: cluster 2026-03-10T10:16:35.770595+0000 mgr.a (mgr.14150) 106 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:38.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:37 vm09 bash[20916]: audit 2026-03-10T10:16:36.730051+0000 mon.a (mon.0) 384 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:38.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:37 vm09 bash[20916]: audit 2026-03-10T10:16:36.730051+0000 mon.a (mon.0) 384 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:38.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:37 vm09 bash[20916]: audit 2026-03-10T10:16:36.732206+0000 mon.a (mon.0) 385 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:38.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:37 vm09 bash[20916]: audit 2026-03-10T10:16:36.732206+0000 mon.a (mon.0) 385 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:38.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:37 vm09 bash[20916]: audit 2026-03-10T10:16:36.732623+0000 mon.a (mon.0) 386 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:38.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:37 vm09 bash[20916]: audit 2026-03-10T10:16:36.732623+0000 mon.a (mon.0) 386 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:38.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:37 vm09 bash[20916]: audit 2026-03-10T10:16:36.735112+0000 mon.a (mon.0) 387 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:38.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:37 vm09 bash[20916]: audit 2026-03-10T10:16:36.735112+0000 mon.a (mon.0) 387 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:38.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:37 vm09 bash[20916]: audit 2026-03-10T10:16:36.739420+0000 mon.a (mon.0) 388 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:38.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:37 vm09 bash[20916]: audit 2026-03-10T10:16:36.739420+0000 mon.a (mon.0) 388 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:38.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:37 vm09 bash[20916]: audit 2026-03-10T10:16:36.743982+0000 mon.a (mon.0) 389 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:38.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:37 vm09 bash[20916]: audit 2026-03-10T10:16:36.743982+0000 mon.a (mon.0) 389 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:38 vm09 bash[20916]: audit 2026-03-10T10:16:37.699656+0000 mon.a (mon.0) 390 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:38 vm09 bash[20916]: audit 2026-03-10T10:16:37.699656+0000 mon.a (mon.0) 390 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:38 vm09 bash[20916]: audit 2026-03-10T10:16:37.705968+0000 mon.a (mon.0) 391 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:38 vm09 bash[20916]: audit 2026-03-10T10:16:37.705968+0000 mon.a (mon.0) 391 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:38 vm09 bash[20916]: cephadm 2026-03-10T10:16:37.707071+0000 mgr.a (mgr.14150) 107 : cephadm [INF] Detected new or changed devices on vm00 2026-03-10T10:16:39.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:38 vm09 bash[20916]: cephadm 2026-03-10T10:16:37.707071+0000 mgr.a (mgr.14150) 107 : cephadm [INF] Detected new or changed devices on vm00 2026-03-10T10:16:39.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:38 vm09 bash[20916]: audit 2026-03-10T10:16:37.711064+0000 mon.a (mon.0) 392 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:39.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:38 vm09 bash[20916]: audit 2026-03-10T10:16:37.711064+0000 mon.a (mon.0) 392 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:39.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:38 vm09 bash[20916]: audit 2026-03-10T10:16:37.713334+0000 mon.a (mon.0) 393 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:39.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:38 vm09 bash[20916]: audit 2026-03-10T10:16:37.713334+0000 mon.a (mon.0) 393 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:39.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:38 vm09 bash[20916]: audit 2026-03-10T10:16:37.713829+0000 mon.a (mon.0) 394 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:39.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:38 vm09 bash[20916]: audit 2026-03-10T10:16:37.713829+0000 mon.a (mon.0) 394 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:39.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:38 vm09 bash[20916]: audit 2026-03-10T10:16:37.716554+0000 mon.a (mon.0) 395 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:38 vm09 bash[20916]: audit 2026-03-10T10:16:37.716554+0000 mon.a (mon.0) 395 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:38 vm09 bash[20916]: audit 2026-03-10T10:16:37.721783+0000 mon.a (mon.0) 396 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:38 vm09 bash[20916]: audit 2026-03-10T10:16:37.721783+0000 mon.a (mon.0) 396 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:38 vm03 bash[21274]: audit 2026-03-10T10:16:37.699656+0000 mon.a (mon.0) 390 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:38 vm03 bash[21274]: audit 2026-03-10T10:16:37.699656+0000 mon.a (mon.0) 390 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:38 vm03 bash[21274]: audit 2026-03-10T10:16:37.705968+0000 mon.a (mon.0) 391 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:38 vm03 bash[21274]: audit 2026-03-10T10:16:37.705968+0000 mon.a (mon.0) 391 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:38 vm03 bash[21274]: cephadm 2026-03-10T10:16:37.707071+0000 mgr.a (mgr.14150) 107 : cephadm [INF] Detected new or changed devices on vm00 2026-03-10T10:16:39.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:38 vm03 bash[21274]: cephadm 2026-03-10T10:16:37.707071+0000 mgr.a (mgr.14150) 107 : cephadm [INF] Detected new or changed devices on vm00 2026-03-10T10:16:39.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:38 vm03 bash[21274]: audit 2026-03-10T10:16:37.711064+0000 mon.a (mon.0) 392 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:39.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:38 vm03 bash[21274]: audit 2026-03-10T10:16:37.711064+0000 mon.a (mon.0) 392 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:39.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:38 vm03 bash[21274]: audit 2026-03-10T10:16:37.713334+0000 mon.a (mon.0) 393 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:39.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:38 vm03 bash[21274]: audit 2026-03-10T10:16:37.713334+0000 mon.a (mon.0) 393 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:39.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:38 vm03 bash[21274]: audit 2026-03-10T10:16:37.713829+0000 mon.a (mon.0) 394 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:39.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:38 vm03 bash[21274]: audit 2026-03-10T10:16:37.713829+0000 mon.a (mon.0) 394 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:39.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:38 vm03 bash[21274]: audit 2026-03-10T10:16:37.716554+0000 mon.a (mon.0) 395 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:38 vm03 bash[21274]: audit 2026-03-10T10:16:37.716554+0000 mon.a (mon.0) 395 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:38 vm03 bash[21274]: audit 2026-03-10T10:16:37.721783+0000 mon.a (mon.0) 396 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:38 vm03 bash[21274]: audit 2026-03-10T10:16:37.721783+0000 mon.a (mon.0) 396 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:38 vm00 bash[20709]: audit 2026-03-10T10:16:37.699656+0000 mon.a (mon.0) 390 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:38 vm00 bash[20709]: audit 2026-03-10T10:16:37.699656+0000 mon.a (mon.0) 390 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:38 vm00 bash[20709]: audit 2026-03-10T10:16:37.705968+0000 mon.a (mon.0) 391 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:38 vm00 bash[20709]: audit 2026-03-10T10:16:37.705968+0000 mon.a (mon.0) 391 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:38 vm00 bash[20709]: cephadm 2026-03-10T10:16:37.707071+0000 mgr.a (mgr.14150) 107 : cephadm [INF] Detected new or changed devices on vm00 2026-03-10T10:16:39.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:38 vm00 bash[20709]: cephadm 2026-03-10T10:16:37.707071+0000 mgr.a (mgr.14150) 107 : cephadm [INF] Detected new or changed devices on vm00 2026-03-10T10:16:39.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:38 vm00 bash[20709]: audit 2026-03-10T10:16:37.711064+0000 mon.a (mon.0) 392 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:39.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:38 vm00 bash[20709]: audit 2026-03-10T10:16:37.711064+0000 mon.a (mon.0) 392 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:39.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:38 vm00 bash[20709]: audit 2026-03-10T10:16:37.713334+0000 mon.a (mon.0) 393 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:39.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:38 vm00 bash[20709]: audit 2026-03-10T10:16:37.713334+0000 mon.a (mon.0) 393 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:39.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:38 vm00 bash[20709]: audit 2026-03-10T10:16:37.713829+0000 mon.a (mon.0) 394 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:39.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:38 vm00 bash[20709]: audit 2026-03-10T10:16:37.713829+0000 mon.a (mon.0) 394 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:39.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:38 vm00 bash[20709]: audit 2026-03-10T10:16:37.716554+0000 mon.a (mon.0) 395 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:38 vm00 bash[20709]: audit 2026-03-10T10:16:37.716554+0000 mon.a (mon.0) 395 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:38 vm00 bash[20709]: audit 2026-03-10T10:16:37.721783+0000 mon.a (mon.0) 396 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:39.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:38 vm00 bash[20709]: audit 2026-03-10T10:16:37.721783+0000 mon.a (mon.0) 396 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:39 vm09 bash[20916]: cluster 2026-03-10T10:16:37.770790+0000 mgr.a (mgr.14150) 108 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:40.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:39 vm09 bash[20916]: cluster 2026-03-10T10:16:37.770790+0000 mgr.a (mgr.14150) 108 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:40.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:39 vm09 bash[20916]: audit 2026-03-10T10:16:38.815840+0000 mon.a (mon.0) 397 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:39 vm09 bash[20916]: audit 2026-03-10T10:16:38.815840+0000 mon.a (mon.0) 397 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:39 vm09 bash[20916]: audit 2026-03-10T10:16:38.820889+0000 mon.a (mon.0) 398 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:39 vm09 bash[20916]: audit 2026-03-10T10:16:38.820889+0000 mon.a (mon.0) 398 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:39 vm09 bash[20916]: audit 2026-03-10T10:16:38.826665+0000 mon.a (mon.0) 399 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:39 vm09 bash[20916]: audit 2026-03-10T10:16:38.826665+0000 mon.a (mon.0) 399 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:39 vm03 bash[21274]: cluster 2026-03-10T10:16:37.770790+0000 mgr.a (mgr.14150) 108 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:40.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:39 vm03 bash[21274]: cluster 2026-03-10T10:16:37.770790+0000 mgr.a (mgr.14150) 108 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:40.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:39 vm03 bash[21274]: audit 2026-03-10T10:16:38.815840+0000 mon.a (mon.0) 397 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:39 vm03 bash[21274]: audit 2026-03-10T10:16:38.815840+0000 mon.a (mon.0) 397 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:39 vm03 bash[21274]: audit 2026-03-10T10:16:38.820889+0000 mon.a (mon.0) 398 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:39 vm03 bash[21274]: audit 2026-03-10T10:16:38.820889+0000 mon.a (mon.0) 398 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:39 vm03 bash[21274]: audit 2026-03-10T10:16:38.826665+0000 mon.a (mon.0) 399 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:39 vm03 bash[21274]: audit 2026-03-10T10:16:38.826665+0000 mon.a (mon.0) 399 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.121 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:39 vm00 bash[20709]: cluster 2026-03-10T10:16:37.770790+0000 mgr.a (mgr.14150) 108 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:40.121 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:39 vm00 bash[20709]: cluster 2026-03-10T10:16:37.770790+0000 mgr.a (mgr.14150) 108 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:40.121 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:39 vm00 bash[20709]: audit 2026-03-10T10:16:38.815840+0000 mon.a (mon.0) 397 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.121 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:39 vm00 bash[20709]: audit 2026-03-10T10:16:38.815840+0000 mon.a (mon.0) 397 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.121 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:39 vm00 bash[20709]: audit 2026-03-10T10:16:38.820889+0000 mon.a (mon.0) 398 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.121 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:39 vm00 bash[20709]: audit 2026-03-10T10:16:38.820889+0000 mon.a (mon.0) 398 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.121 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:39 vm00 bash[20709]: audit 2026-03-10T10:16:38.826665+0000 mon.a (mon.0) 399 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:40.121 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:39 vm00 bash[20709]: audit 2026-03-10T10:16:38.826665+0000 mon.a (mon.0) 399 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:41.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:40 vm09 bash[20916]: audit 2026-03-10T10:16:40.123008+0000 mon.a (mon.0) 400 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T10:16:41.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:40 vm09 bash[20916]: audit 2026-03-10T10:16:40.123008+0000 mon.a (mon.0) 400 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T10:16:41.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:40 vm03 bash[21274]: audit 2026-03-10T10:16:40.123008+0000 mon.a (mon.0) 400 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T10:16:41.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:40 vm03 bash[21274]: audit 2026-03-10T10:16:40.123008+0000 mon.a (mon.0) 400 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T10:16:41.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:40 vm00 bash[20709]: audit 2026-03-10T10:16:40.123008+0000 mon.a (mon.0) 400 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T10:16:41.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:40 vm00 bash[20709]: audit 2026-03-10T10:16:40.123008+0000 mon.a (mon.0) 400 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T10:16:42.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:41 vm09 bash[20916]: cluster 2026-03-10T10:16:39.771053+0000 mgr.a (mgr.14150) 109 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:42.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:41 vm09 bash[20916]: cluster 2026-03-10T10:16:39.771053+0000 mgr.a (mgr.14150) 109 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:42.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:41 vm09 bash[20916]: audit 2026-03-10T10:16:40.708358+0000 mon.a (mon.0) 401 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T10:16:42.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:41 vm09 bash[20916]: audit 2026-03-10T10:16:40.708358+0000 mon.a (mon.0) 401 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T10:16:42.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:41 vm09 bash[20916]: cluster 2026-03-10T10:16:40.709739+0000 mon.a (mon.0) 402 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-10T10:16:42.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:41 vm09 bash[20916]: cluster 2026-03-10T10:16:40.709739+0000 mon.a (mon.0) 402 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-10T10:16:42.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:41 vm09 bash[20916]: audit 2026-03-10T10:16:40.709871+0000 mon.a (mon.0) 403 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]: dispatch 2026-03-10T10:16:42.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:41 vm09 bash[20916]: audit 2026-03-10T10:16:40.709871+0000 mon.a (mon.0) 403 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]: dispatch 2026-03-10T10:16:42.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:41 vm09 bash[20916]: audit 2026-03-10T10:16:40.709958+0000 mon.a (mon.0) 404 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:42.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:41 vm09 bash[20916]: audit 2026-03-10T10:16:40.709958+0000 mon.a (mon.0) 404 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:42.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:41 vm03 bash[21274]: cluster 2026-03-10T10:16:39.771053+0000 mgr.a (mgr.14150) 109 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:42.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:41 vm03 bash[21274]: cluster 2026-03-10T10:16:39.771053+0000 mgr.a (mgr.14150) 109 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:42.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:41 vm03 bash[21274]: audit 2026-03-10T10:16:40.708358+0000 mon.a (mon.0) 401 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T10:16:42.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:41 vm03 bash[21274]: audit 2026-03-10T10:16:40.708358+0000 mon.a (mon.0) 401 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T10:16:42.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:41 vm03 bash[21274]: cluster 2026-03-10T10:16:40.709739+0000 mon.a (mon.0) 402 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-10T10:16:42.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:41 vm03 bash[21274]: cluster 2026-03-10T10:16:40.709739+0000 mon.a (mon.0) 402 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-10T10:16:42.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:41 vm03 bash[21274]: audit 2026-03-10T10:16:40.709871+0000 mon.a (mon.0) 403 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]: dispatch 2026-03-10T10:16:42.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:41 vm03 bash[21274]: audit 2026-03-10T10:16:40.709871+0000 mon.a (mon.0) 403 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]: dispatch 2026-03-10T10:16:42.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:41 vm03 bash[21274]: audit 2026-03-10T10:16:40.709958+0000 mon.a (mon.0) 404 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:42.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:41 vm03 bash[21274]: audit 2026-03-10T10:16:40.709958+0000 mon.a (mon.0) 404 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:42.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:41 vm00 bash[20709]: cluster 2026-03-10T10:16:39.771053+0000 mgr.a (mgr.14150) 109 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:42.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:41 vm00 bash[20709]: cluster 2026-03-10T10:16:39.771053+0000 mgr.a (mgr.14150) 109 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:42.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:41 vm00 bash[20709]: audit 2026-03-10T10:16:40.708358+0000 mon.a (mon.0) 401 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T10:16:42.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:41 vm00 bash[20709]: audit 2026-03-10T10:16:40.708358+0000 mon.a (mon.0) 401 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T10:16:42.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:41 vm00 bash[20709]: cluster 2026-03-10T10:16:40.709739+0000 mon.a (mon.0) 402 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-10T10:16:42.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:41 vm00 bash[20709]: cluster 2026-03-10T10:16:40.709739+0000 mon.a (mon.0) 402 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-10T10:16:42.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:41 vm00 bash[20709]: audit 2026-03-10T10:16:40.709871+0000 mon.a (mon.0) 403 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]: dispatch 2026-03-10T10:16:42.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:41 vm00 bash[20709]: audit 2026-03-10T10:16:40.709871+0000 mon.a (mon.0) 403 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]: dispatch 2026-03-10T10:16:42.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:41 vm00 bash[20709]: audit 2026-03-10T10:16:40.709958+0000 mon.a (mon.0) 404 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:42.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:41 vm00 bash[20709]: audit 2026-03-10T10:16:40.709958+0000 mon.a (mon.0) 404 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:42.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:42 vm00 bash[20709]: audit 2026-03-10T10:16:41.711063+0000 mon.a (mon.0) 405 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]': finished 2026-03-10T10:16:42.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:42 vm00 bash[20709]: audit 2026-03-10T10:16:41.711063+0000 mon.a (mon.0) 405 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]': finished 2026-03-10T10:16:42.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:42 vm00 bash[20709]: cluster 2026-03-10T10:16:41.713549+0000 mon.a (mon.0) 406 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-10T10:16:42.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:42 vm00 bash[20709]: cluster 2026-03-10T10:16:41.713549+0000 mon.a (mon.0) 406 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-10T10:16:42.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:42 vm00 bash[20709]: audit 2026-03-10T10:16:41.714429+0000 mon.a (mon.0) 407 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:42.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:42 vm00 bash[20709]: audit 2026-03-10T10:16:41.714429+0000 mon.a (mon.0) 407 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:42.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:42 vm00 bash[20709]: audit 2026-03-10T10:16:41.720159+0000 mon.a (mon.0) 408 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:42.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:42 vm00 bash[20709]: audit 2026-03-10T10:16:41.720159+0000 mon.a (mon.0) 408 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:43.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:42 vm09 bash[20916]: audit 2026-03-10T10:16:41.711063+0000 mon.a (mon.0) 405 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]': finished 2026-03-10T10:16:43.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:42 vm09 bash[20916]: audit 2026-03-10T10:16:41.711063+0000 mon.a (mon.0) 405 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]': finished 2026-03-10T10:16:43.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:42 vm09 bash[20916]: cluster 2026-03-10T10:16:41.713549+0000 mon.a (mon.0) 406 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-10T10:16:43.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:42 vm09 bash[20916]: cluster 2026-03-10T10:16:41.713549+0000 mon.a (mon.0) 406 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-10T10:16:43.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:42 vm09 bash[20916]: audit 2026-03-10T10:16:41.714429+0000 mon.a (mon.0) 407 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:43.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:42 vm09 bash[20916]: audit 2026-03-10T10:16:41.714429+0000 mon.a (mon.0) 407 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:43.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:42 vm09 bash[20916]: audit 2026-03-10T10:16:41.720159+0000 mon.a (mon.0) 408 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:43.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:42 vm09 bash[20916]: audit 2026-03-10T10:16:41.720159+0000 mon.a (mon.0) 408 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:43.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:42 vm03 bash[21274]: audit 2026-03-10T10:16:41.711063+0000 mon.a (mon.0) 405 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]': finished 2026-03-10T10:16:43.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:42 vm03 bash[21274]: audit 2026-03-10T10:16:41.711063+0000 mon.a (mon.0) 405 : audit [INF] from='osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]': finished 2026-03-10T10:16:43.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:42 vm03 bash[21274]: cluster 2026-03-10T10:16:41.713549+0000 mon.a (mon.0) 406 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-10T10:16:43.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:42 vm03 bash[21274]: cluster 2026-03-10T10:16:41.713549+0000 mon.a (mon.0) 406 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-10T10:16:43.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:42 vm03 bash[21274]: audit 2026-03-10T10:16:41.714429+0000 mon.a (mon.0) 407 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:43.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:42 vm03 bash[21274]: audit 2026-03-10T10:16:41.714429+0000 mon.a (mon.0) 407 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:43.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:42 vm03 bash[21274]: audit 2026-03-10T10:16:41.720159+0000 mon.a (mon.0) 408 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:43.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:42 vm03 bash[21274]: audit 2026-03-10T10:16:41.720159+0000 mon.a (mon.0) 408 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:43.729 INFO:teuthology.orchestra.run.vm00.stdout:Created osd(s) 0 on host 'vm00' 2026-03-10T10:16:43.814 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: cluster 2026-03-10T10:16:41.111318+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: cluster 2026-03-10T10:16:41.111318+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: cluster 2026-03-10T10:16:41.111364+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: cluster 2026-03-10T10:16:41.111364+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: cluster 2026-03-10T10:16:41.771272+0000 mgr.a (mgr.14150) 110 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: cluster 2026-03-10T10:16:41.771272+0000 mgr.a (mgr.14150) 110 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: audit 2026-03-10T10:16:42.717582+0000 mon.a (mon.0) 409 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: audit 2026-03-10T10:16:42.717582+0000 mon.a (mon.0) 409 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: cluster 2026-03-10T10:16:42.722457+0000 mon.a (mon.0) 410 : cluster [INF] osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291] boot 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: cluster 2026-03-10T10:16:42.722457+0000 mon.a (mon.0) 410 : cluster [INF] osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291] boot 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: cluster 2026-03-10T10:16:42.722499+0000 mon.a (mon.0) 411 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: cluster 2026-03-10T10:16:42.722499+0000 mon.a (mon.0) 411 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: audit 2026-03-10T10:16:42.724259+0000 mon.a (mon.0) 412 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: audit 2026-03-10T10:16:42.724259+0000 mon.a (mon.0) 412 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: audit 2026-03-10T10:16:43.711888+0000 mon.a (mon.0) 413 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: audit 2026-03-10T10:16:43.711888+0000 mon.a (mon.0) 413 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: audit 2026-03-10T10:16:43.713942+0000 mon.a (mon.0) 414 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: audit 2026-03-10T10:16:43.713942+0000 mon.a (mon.0) 414 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: audit 2026-03-10T10:16:43.717610+0000 mon.a (mon.0) 415 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: audit 2026-03-10T10:16:43.717610+0000 mon.a (mon.0) 415 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: audit 2026-03-10T10:16:43.720867+0000 mon.a (mon.0) 416 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: audit 2026-03-10T10:16:43.720867+0000 mon.a (mon.0) 416 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: audit 2026-03-10T10:16:43.721347+0000 mon.a (mon.0) 417 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: audit 2026-03-10T10:16:43.721347+0000 mon.a (mon.0) 417 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: audit 2026-03-10T10:16:43.724715+0000 mon.a (mon.0) 418 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:43.815 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:43 vm00 bash[20709]: audit 2026-03-10T10:16:43.724715+0000 mon.a (mon.0) 418 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:43.815 DEBUG:teuthology.orchestra.run.vm00:osd.0> sudo journalctl -f -n 0 -u ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@osd.0.service 2026-03-10T10:16:43.816 INFO:tasks.cephadm:Deploying osd.1 on vm03 with /dev/vde... 2026-03-10T10:16:43.816 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- lvm zap /dev/vde 2026-03-10T10:16:44.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: cluster 2026-03-10T10:16:41.111318+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:16:44.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: cluster 2026-03-10T10:16:41.111318+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:16:44.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: cluster 2026-03-10T10:16:41.111364+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:16:44.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: cluster 2026-03-10T10:16:41.111364+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:16:44.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: cluster 2026-03-10T10:16:41.771272+0000 mgr.a (mgr.14150) 110 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:44.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: cluster 2026-03-10T10:16:41.771272+0000 mgr.a (mgr.14150) 110 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:44.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: audit 2026-03-10T10:16:42.717582+0000 mon.a (mon.0) 409 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:44.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: audit 2026-03-10T10:16:42.717582+0000 mon.a (mon.0) 409 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:44.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: cluster 2026-03-10T10:16:42.722457+0000 mon.a (mon.0) 410 : cluster [INF] osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291] boot 2026-03-10T10:16:44.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: cluster 2026-03-10T10:16:42.722457+0000 mon.a (mon.0) 410 : cluster [INF] osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291] boot 2026-03-10T10:16:44.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: cluster 2026-03-10T10:16:42.722499+0000 mon.a (mon.0) 411 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-10T10:16:44.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: cluster 2026-03-10T10:16:42.722499+0000 mon.a (mon.0) 411 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-10T10:16:44.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: audit 2026-03-10T10:16:42.724259+0000 mon.a (mon.0) 412 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:44.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: audit 2026-03-10T10:16:42.724259+0000 mon.a (mon.0) 412 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:44.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: audit 2026-03-10T10:16:43.711888+0000 mon.a (mon.0) 413 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:44.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: audit 2026-03-10T10:16:43.711888+0000 mon.a (mon.0) 413 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:44.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: audit 2026-03-10T10:16:43.713942+0000 mon.a (mon.0) 414 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:16:44.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: audit 2026-03-10T10:16:43.713942+0000 mon.a (mon.0) 414 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:16:44.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: audit 2026-03-10T10:16:43.717610+0000 mon.a (mon.0) 415 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:44.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: audit 2026-03-10T10:16:43.717610+0000 mon.a (mon.0) 415 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:44.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: audit 2026-03-10T10:16:43.720867+0000 mon.a (mon.0) 416 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:44.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: audit 2026-03-10T10:16:43.720867+0000 mon.a (mon.0) 416 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:44.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: audit 2026-03-10T10:16:43.721347+0000 mon.a (mon.0) 417 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:44.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: audit 2026-03-10T10:16:43.721347+0000 mon.a (mon.0) 417 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:44.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: audit 2026-03-10T10:16:43.724715+0000 mon.a (mon.0) 418 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:44.005 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:43 vm09 bash[20916]: audit 2026-03-10T10:16:43.724715+0000 mon.a (mon.0) 418 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: cluster 2026-03-10T10:16:41.111318+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: cluster 2026-03-10T10:16:41.111318+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: cluster 2026-03-10T10:16:41.111364+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: cluster 2026-03-10T10:16:41.111364+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: cluster 2026-03-10T10:16:41.771272+0000 mgr.a (mgr.14150) 110 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: cluster 2026-03-10T10:16:41.771272+0000 mgr.a (mgr.14150) 110 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: audit 2026-03-10T10:16:42.717582+0000 mon.a (mon.0) 409 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: audit 2026-03-10T10:16:42.717582+0000 mon.a (mon.0) 409 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: cluster 2026-03-10T10:16:42.722457+0000 mon.a (mon.0) 410 : cluster [INF] osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291] boot 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: cluster 2026-03-10T10:16:42.722457+0000 mon.a (mon.0) 410 : cluster [INF] osd.0 [v2:192.168.123.100:6802/2417190291,v1:192.168.123.100:6803/2417190291] boot 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: cluster 2026-03-10T10:16:42.722499+0000 mon.a (mon.0) 411 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: cluster 2026-03-10T10:16:42.722499+0000 mon.a (mon.0) 411 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: audit 2026-03-10T10:16:42.724259+0000 mon.a (mon.0) 412 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: audit 2026-03-10T10:16:42.724259+0000 mon.a (mon.0) 412 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: audit 2026-03-10T10:16:43.711888+0000 mon.a (mon.0) 413 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: audit 2026-03-10T10:16:43.711888+0000 mon.a (mon.0) 413 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: audit 2026-03-10T10:16:43.713942+0000 mon.a (mon.0) 414 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: audit 2026-03-10T10:16:43.713942+0000 mon.a (mon.0) 414 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: audit 2026-03-10T10:16:43.717610+0000 mon.a (mon.0) 415 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: audit 2026-03-10T10:16:43.717610+0000 mon.a (mon.0) 415 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: audit 2026-03-10T10:16:43.720867+0000 mon.a (mon.0) 416 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: audit 2026-03-10T10:16:43.720867+0000 mon.a (mon.0) 416 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: audit 2026-03-10T10:16:43.721347+0000 mon.a (mon.0) 417 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: audit 2026-03-10T10:16:43.721347+0000 mon.a (mon.0) 417 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: audit 2026-03-10T10:16:43.724715+0000 mon.a (mon.0) 418 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:43 vm03 bash[21274]: audit 2026-03-10T10:16:43.724715+0000 mon.a (mon.0) 418 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:45.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:44 vm09 bash[20916]: audit 2026-03-10T10:16:43.731340+0000 mon.a (mon.0) 419 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:45.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:44 vm09 bash[20916]: audit 2026-03-10T10:16:43.731340+0000 mon.a (mon.0) 419 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:45.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:44 vm09 bash[20916]: cluster 2026-03-10T10:16:43.749033+0000 mon.a (mon.0) 420 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-10T10:16:45.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:44 vm09 bash[20916]: cluster 2026-03-10T10:16:43.749033+0000 mon.a (mon.0) 420 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-10T10:16:45.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:44 vm09 bash[20916]: cluster 2026-03-10T10:16:43.771471+0000 mgr.a (mgr.14150) 111 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:45.004 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:44 vm09 bash[20916]: cluster 2026-03-10T10:16:43.771471+0000 mgr.a (mgr.14150) 111 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:45.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:44 vm03 bash[21274]: audit 2026-03-10T10:16:43.731340+0000 mon.a (mon.0) 419 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:45.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:44 vm03 bash[21274]: audit 2026-03-10T10:16:43.731340+0000 mon.a (mon.0) 419 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:45.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:44 vm03 bash[21274]: cluster 2026-03-10T10:16:43.749033+0000 mon.a (mon.0) 420 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-10T10:16:45.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:44 vm03 bash[21274]: cluster 2026-03-10T10:16:43.749033+0000 mon.a (mon.0) 420 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-10T10:16:45.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:44 vm03 bash[21274]: cluster 2026-03-10T10:16:43.771471+0000 mgr.a (mgr.14150) 111 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:45.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:44 vm03 bash[21274]: cluster 2026-03-10T10:16:43.771471+0000 mgr.a (mgr.14150) 111 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:45.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:44 vm00 bash[20709]: audit 2026-03-10T10:16:43.731340+0000 mon.a (mon.0) 419 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:45.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:44 vm00 bash[20709]: audit 2026-03-10T10:16:43.731340+0000 mon.a (mon.0) 419 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:45.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:44 vm00 bash[20709]: cluster 2026-03-10T10:16:43.749033+0000 mon.a (mon.0) 420 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-10T10:16:45.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:44 vm00 bash[20709]: cluster 2026-03-10T10:16:43.749033+0000 mon.a (mon.0) 420 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-10T10:16:45.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:44 vm00 bash[20709]: cluster 2026-03-10T10:16:43.771471+0000 mgr.a (mgr.14150) 111 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:45.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:44 vm00 bash[20709]: cluster 2026-03-10T10:16:43.771471+0000 mgr.a (mgr.14150) 111 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:47.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:46 vm03 bash[21274]: cluster 2026-03-10T10:16:45.771696+0000 mgr.a (mgr.14150) 112 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:47.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:46 vm03 bash[21274]: cluster 2026-03-10T10:16:45.771696+0000 mgr.a (mgr.14150) 112 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:47.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:46 vm00 bash[20709]: cluster 2026-03-10T10:16:45.771696+0000 mgr.a (mgr.14150) 112 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:47.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:46 vm00 bash[20709]: cluster 2026-03-10T10:16:45.771696+0000 mgr.a (mgr.14150) 112 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:47.254 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:46 vm09 bash[20916]: cluster 2026-03-10T10:16:45.771696+0000 mgr.a (mgr.14150) 112 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:47.254 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:46 vm09 bash[20916]: cluster 2026-03-10T10:16:45.771696+0000 mgr.a (mgr.14150) 112 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:48.430 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.b/config 2026-03-10T10:16:49.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:48 vm00 bash[20709]: cluster 2026-03-10T10:16:47.771911+0000 mgr.a (mgr.14150) 113 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:49.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:48 vm00 bash[20709]: cluster 2026-03-10T10:16:47.771911+0000 mgr.a (mgr.14150) 113 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:49.221 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:48 vm03 bash[21274]: cluster 2026-03-10T10:16:47.771911+0000 mgr.a (mgr.14150) 113 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:49.221 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:48 vm03 bash[21274]: cluster 2026-03-10T10:16:47.771911+0000 mgr.a (mgr.14150) 113 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:49.249 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-10T10:16:49.254 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:48 vm09 bash[20916]: cluster 2026-03-10T10:16:47.771911+0000 mgr.a (mgr.14150) 113 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:49.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:48 vm09 bash[20916]: cluster 2026-03-10T10:16:47.771911+0000 mgr.a (mgr.14150) 113 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:49.265 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph orch daemon add osd vm03:/dev/vde 2026-03-10T10:16:51.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:50 vm00 bash[20709]: cluster 2026-03-10T10:16:49.772148+0000 mgr.a (mgr.14150) 114 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:51.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:50 vm00 bash[20709]: cluster 2026-03-10T10:16:49.772148+0000 mgr.a (mgr.14150) 114 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:51.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:50 vm09 bash[20916]: cluster 2026-03-10T10:16:49.772148+0000 mgr.a (mgr.14150) 114 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:51.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:50 vm09 bash[20916]: cluster 2026-03-10T10:16:49.772148+0000 mgr.a (mgr.14150) 114 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:51.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:50 vm03 bash[21274]: cluster 2026-03-10T10:16:49.772148+0000 mgr.a (mgr.14150) 114 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:51.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:50 vm03 bash[21274]: cluster 2026-03-10T10:16:49.772148+0000 mgr.a (mgr.14150) 114 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:53.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:52 vm03 bash[21274]: cluster 2026-03-10T10:16:51.772364+0000 mgr.a (mgr.14150) 115 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:53.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:52 vm03 bash[21274]: cluster 2026-03-10T10:16:51.772364+0000 mgr.a (mgr.14150) 115 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:53.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:52 vm00 bash[20709]: cluster 2026-03-10T10:16:51.772364+0000 mgr.a (mgr.14150) 115 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:53.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:52 vm00 bash[20709]: cluster 2026-03-10T10:16:51.772364+0000 mgr.a (mgr.14150) 115 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:52 vm09 bash[20916]: cluster 2026-03-10T10:16:51.772364+0000 mgr.a (mgr.14150) 115 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:53.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:52 vm09 bash[20916]: cluster 2026-03-10T10:16:51.772364+0000 mgr.a (mgr.14150) 115 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:53.877 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.b/config 2026-03-10T10:16:55.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:54 vm00 bash[20709]: cluster 2026-03-10T10:16:53.772584+0000 mgr.a (mgr.14150) 116 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:55.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:54 vm00 bash[20709]: cluster 2026-03-10T10:16:53.772584+0000 mgr.a (mgr.14150) 116 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:55.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:54 vm00 bash[20709]: audit 2026-03-10T10:16:54.138141+0000 mgr.a (mgr.14150) 117 : audit [DBG] from='client.24131 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm03:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:55.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:54 vm00 bash[20709]: audit 2026-03-10T10:16:54.138141+0000 mgr.a (mgr.14150) 117 : audit [DBG] from='client.24131 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm03:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:55.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:54 vm00 bash[20709]: audit 2026-03-10T10:16:54.139665+0000 mon.a (mon.0) 421 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:16:55.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:54 vm00 bash[20709]: audit 2026-03-10T10:16:54.139665+0000 mon.a (mon.0) 421 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:16:55.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:54 vm00 bash[20709]: audit 2026-03-10T10:16:54.140827+0000 mon.a (mon.0) 422 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:16:55.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:54 vm00 bash[20709]: audit 2026-03-10T10:16:54.140827+0000 mon.a (mon.0) 422 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:16:55.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:54 vm00 bash[20709]: audit 2026-03-10T10:16:54.141182+0000 mon.a (mon.0) 423 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:55.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:54 vm00 bash[20709]: audit 2026-03-10T10:16:54.141182+0000 mon.a (mon.0) 423 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:55.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:54 vm09 bash[20916]: cluster 2026-03-10T10:16:53.772584+0000 mgr.a (mgr.14150) 116 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:55.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:54 vm09 bash[20916]: cluster 2026-03-10T10:16:53.772584+0000 mgr.a (mgr.14150) 116 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:55.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:54 vm09 bash[20916]: audit 2026-03-10T10:16:54.138141+0000 mgr.a (mgr.14150) 117 : audit [DBG] from='client.24131 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm03:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:55.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:54 vm09 bash[20916]: audit 2026-03-10T10:16:54.138141+0000 mgr.a (mgr.14150) 117 : audit [DBG] from='client.24131 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm03:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:55.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:54 vm09 bash[20916]: audit 2026-03-10T10:16:54.139665+0000 mon.a (mon.0) 421 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:16:55.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:54 vm09 bash[20916]: audit 2026-03-10T10:16:54.139665+0000 mon.a (mon.0) 421 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:16:55.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:54 vm09 bash[20916]: audit 2026-03-10T10:16:54.140827+0000 mon.a (mon.0) 422 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:16:55.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:54 vm09 bash[20916]: audit 2026-03-10T10:16:54.140827+0000 mon.a (mon.0) 422 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:16:55.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:54 vm09 bash[20916]: audit 2026-03-10T10:16:54.141182+0000 mon.a (mon.0) 423 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:55.255 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:54 vm09 bash[20916]: audit 2026-03-10T10:16:54.141182+0000 mon.a (mon.0) 423 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:55.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:54 vm03 bash[21274]: cluster 2026-03-10T10:16:53.772584+0000 mgr.a (mgr.14150) 116 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:55.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:54 vm03 bash[21274]: cluster 2026-03-10T10:16:53.772584+0000 mgr.a (mgr.14150) 116 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:55.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:54 vm03 bash[21274]: audit 2026-03-10T10:16:54.138141+0000 mgr.a (mgr.14150) 117 : audit [DBG] from='client.24131 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm03:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:55.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:54 vm03 bash[21274]: audit 2026-03-10T10:16:54.138141+0000 mgr.a (mgr.14150) 117 : audit [DBG] from='client.24131 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm03:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:16:55.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:54 vm03 bash[21274]: audit 2026-03-10T10:16:54.139665+0000 mon.a (mon.0) 421 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:16:55.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:54 vm03 bash[21274]: audit 2026-03-10T10:16:54.139665+0000 mon.a (mon.0) 421 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:16:55.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:54 vm03 bash[21274]: audit 2026-03-10T10:16:54.140827+0000 mon.a (mon.0) 422 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:16:55.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:54 vm03 bash[21274]: audit 2026-03-10T10:16:54.140827+0000 mon.a (mon.0) 422 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:16:55.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:54 vm03 bash[21274]: audit 2026-03-10T10:16:54.141182+0000 mon.a (mon.0) 423 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:55.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:54 vm03 bash[21274]: audit 2026-03-10T10:16:54.141182+0000 mon.a (mon.0) 423 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:57.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:57 vm00 bash[20709]: cluster 2026-03-10T10:16:55.772785+0000 mgr.a (mgr.14150) 118 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:57.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:57 vm00 bash[20709]: cluster 2026-03-10T10:16:55.772785+0000 mgr.a (mgr.14150) 118 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:57.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:57 vm00 bash[20709]: audit 2026-03-10T10:16:56.335074+0000 mon.a (mon.0) 424 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:57 vm00 bash[20709]: audit 2026-03-10T10:16:56.335074+0000 mon.a (mon.0) 424 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:57 vm00 bash[20709]: audit 2026-03-10T10:16:56.342367+0000 mon.a (mon.0) 425 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:57 vm00 bash[20709]: audit 2026-03-10T10:16:56.342367+0000 mon.a (mon.0) 425 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:57 vm00 bash[20709]: audit 2026-03-10T10:16:56.346569+0000 mon.a (mon.0) 426 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:57 vm00 bash[20709]: audit 2026-03-10T10:16:56.346569+0000 mon.a (mon.0) 426 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:57 vm00 bash[20709]: audit 2026-03-10T10:16:56.776864+0000 mon.a (mon.0) 427 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:57 vm00 bash[20709]: audit 2026-03-10T10:16:56.776864+0000 mon.a (mon.0) 427 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:57 vm00 bash[20709]: audit 2026-03-10T10:16:56.781059+0000 mon.a (mon.0) 428 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:57 vm00 bash[20709]: audit 2026-03-10T10:16:56.781059+0000 mon.a (mon.0) 428 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:57 vm00 bash[20709]: audit 2026-03-10T10:16:56.786135+0000 mon.a (mon.0) 429 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:57 vm00 bash[20709]: audit 2026-03-10T10:16:56.786135+0000 mon.a (mon.0) 429 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:57 vm09 bash[20916]: cluster 2026-03-10T10:16:55.772785+0000 mgr.a (mgr.14150) 118 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:57.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:57 vm09 bash[20916]: cluster 2026-03-10T10:16:55.772785+0000 mgr.a (mgr.14150) 118 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:57.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:57 vm09 bash[20916]: audit 2026-03-10T10:16:56.335074+0000 mon.a (mon.0) 424 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:57 vm09 bash[20916]: audit 2026-03-10T10:16:56.335074+0000 mon.a (mon.0) 424 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:57 vm09 bash[20916]: audit 2026-03-10T10:16:56.342367+0000 mon.a (mon.0) 425 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:57 vm09 bash[20916]: audit 2026-03-10T10:16:56.342367+0000 mon.a (mon.0) 425 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:57 vm09 bash[20916]: audit 2026-03-10T10:16:56.346569+0000 mon.a (mon.0) 426 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:57 vm09 bash[20916]: audit 2026-03-10T10:16:56.346569+0000 mon.a (mon.0) 426 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:57 vm09 bash[20916]: audit 2026-03-10T10:16:56.776864+0000 mon.a (mon.0) 427 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:57 vm09 bash[20916]: audit 2026-03-10T10:16:56.776864+0000 mon.a (mon.0) 427 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:57 vm09 bash[20916]: audit 2026-03-10T10:16:56.781059+0000 mon.a (mon.0) 428 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:57 vm09 bash[20916]: audit 2026-03-10T10:16:56.781059+0000 mon.a (mon.0) 428 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:57 vm09 bash[20916]: audit 2026-03-10T10:16:56.786135+0000 mon.a (mon.0) 429 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:57 vm09 bash[20916]: audit 2026-03-10T10:16:56.786135+0000 mon.a (mon.0) 429 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:57 vm03 bash[21274]: cluster 2026-03-10T10:16:55.772785+0000 mgr.a (mgr.14150) 118 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:57.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:57 vm03 bash[21274]: cluster 2026-03-10T10:16:55.772785+0000 mgr.a (mgr.14150) 118 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:57.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:57 vm03 bash[21274]: audit 2026-03-10T10:16:56.335074+0000 mon.a (mon.0) 424 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:57 vm03 bash[21274]: audit 2026-03-10T10:16:56.335074+0000 mon.a (mon.0) 424 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:57 vm03 bash[21274]: audit 2026-03-10T10:16:56.342367+0000 mon.a (mon.0) 425 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:57 vm03 bash[21274]: audit 2026-03-10T10:16:56.342367+0000 mon.a (mon.0) 425 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:57 vm03 bash[21274]: audit 2026-03-10T10:16:56.346569+0000 mon.a (mon.0) 426 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:57 vm03 bash[21274]: audit 2026-03-10T10:16:56.346569+0000 mon.a (mon.0) 426 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:57 vm03 bash[21274]: audit 2026-03-10T10:16:56.776864+0000 mon.a (mon.0) 427 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:57 vm03 bash[21274]: audit 2026-03-10T10:16:56.776864+0000 mon.a (mon.0) 427 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:57 vm03 bash[21274]: audit 2026-03-10T10:16:56.781059+0000 mon.a (mon.0) 428 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:57 vm03 bash[21274]: audit 2026-03-10T10:16:56.781059+0000 mon.a (mon.0) 428 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:57 vm03 bash[21274]: audit 2026-03-10T10:16:56.786135+0000 mon.a (mon.0) 429 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:57.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:57 vm03 bash[21274]: audit 2026-03-10T10:16:56.786135+0000 mon.a (mon.0) 429 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: cluster 2026-03-10T10:16:57.773000+0000 mgr.a (mgr.14150) 119 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: cluster 2026-03-10T10:16:57.773000+0000 mgr.a (mgr.14150) 119 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:58.152365+0000 mon.a (mon.0) 430 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:58.152365+0000 mon.a (mon.0) 430 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:58.157627+0000 mon.a (mon.0) 431 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:58.157627+0000 mon.a (mon.0) 431 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:58.159442+0000 mon.a (mon.0) 432 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:58.159442+0000 mon.a (mon.0) 432 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:58.161472+0000 mon.a (mon.0) 433 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:58.161472+0000 mon.a (mon.0) 433 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:58.162018+0000 mon.a (mon.0) 434 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:58.162018+0000 mon.a (mon.0) 434 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:58.164699+0000 mon.a (mon.0) 435 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:58.164699+0000 mon.a (mon.0) 435 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:58.170614+0000 mon.a (mon.0) 436 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:58.170614+0000 mon.a (mon.0) 436 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:59.079929+0000 mon.a (mon.0) 437 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:59.079929+0000 mon.a (mon.0) 437 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:59.101683+0000 mon.a (mon.0) 438 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:59.101683+0000 mon.a (mon.0) 438 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:59.110813+0000 mon.a (mon.0) 439 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:16:59 vm00 bash[20709]: audit 2026-03-10T10:16:59.110813+0000 mon.a (mon.0) 439 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.505 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: cluster 2026-03-10T10:16:57.773000+0000 mgr.a (mgr.14150) 119 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:59.505 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: cluster 2026-03-10T10:16:57.773000+0000 mgr.a (mgr.14150) 119 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:59.505 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:58.152365+0000 mon.a (mon.0) 430 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:58.152365+0000 mon.a (mon.0) 430 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:58.157627+0000 mon.a (mon.0) 431 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:58.157627+0000 mon.a (mon.0) 431 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:58.159442+0000 mon.a (mon.0) 432 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:58.159442+0000 mon.a (mon.0) 432 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:58.161472+0000 mon.a (mon.0) 433 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:58.161472+0000 mon.a (mon.0) 433 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:58.162018+0000 mon.a (mon.0) 434 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:58.162018+0000 mon.a (mon.0) 434 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:58.164699+0000 mon.a (mon.0) 435 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:58.164699+0000 mon.a (mon.0) 435 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:58.170614+0000 mon.a (mon.0) 436 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:58.170614+0000 mon.a (mon.0) 436 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:59.079929+0000 mon.a (mon.0) 437 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:59.079929+0000 mon.a (mon.0) 437 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:59.101683+0000 mon.a (mon.0) 438 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:59.101683+0000 mon.a (mon.0) 438 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:59.110813+0000 mon.a (mon.0) 439 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:16:59 vm09 bash[20916]: audit 2026-03-10T10:16:59.110813+0000 mon.a (mon.0) 439 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: cluster 2026-03-10T10:16:57.773000+0000 mgr.a (mgr.14150) 119 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: cluster 2026-03-10T10:16:57.773000+0000 mgr.a (mgr.14150) 119 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:16:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:58.152365+0000 mon.a (mon.0) 430 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:58.152365+0000 mon.a (mon.0) 430 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:58.157627+0000 mon.a (mon.0) 431 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:58.157627+0000 mon.a (mon.0) 431 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:58.159442+0000 mon.a (mon.0) 432 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:58.159442+0000 mon.a (mon.0) 432 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:16:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:58.161472+0000 mon.a (mon.0) 433 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:58.161472+0000 mon.a (mon.0) 433 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:16:59.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:58.162018+0000 mon.a (mon.0) 434 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:59.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:58.162018+0000 mon.a (mon.0) 434 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:16:59.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:58.164699+0000 mon.a (mon.0) 435 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:58.164699+0000 mon.a (mon.0) 435 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:58.170614+0000 mon.a (mon.0) 436 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:58.170614+0000 mon.a (mon.0) 436 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:59.079929+0000 mon.a (mon.0) 437 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:59.079929+0000 mon.a (mon.0) 437 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:59.101683+0000 mon.a (mon.0) 438 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:59.101683+0000 mon.a (mon.0) 438 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:59.110813+0000 mon.a (mon.0) 439 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:16:59.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:16:59 vm03 bash[21274]: audit 2026-03-10T10:16:59.110813+0000 mon.a (mon.0) 439 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:00.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:00 vm03 bash[21274]: audit 2026-03-10T10:16:59.706040+0000 mon.a (mon.0) 440 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]: dispatch 2026-03-10T10:17:00.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:00 vm03 bash[21274]: audit 2026-03-10T10:16:59.706040+0000 mon.a (mon.0) 440 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]: dispatch 2026-03-10T10:17:00.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:00 vm03 bash[21274]: audit 2026-03-10T10:16:59.706453+0000 mon.b (mon.2) 3 : audit [INF] from='client.? 192.168.123.103:0/1012162837' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]: dispatch 2026-03-10T10:17:00.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:00 vm03 bash[21274]: audit 2026-03-10T10:16:59.706453+0000 mon.b (mon.2) 3 : audit [INF] from='client.? 192.168.123.103:0/1012162837' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]: dispatch 2026-03-10T10:17:00.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:00 vm03 bash[21274]: audit 2026-03-10T10:16:59.709063+0000 mon.a (mon.0) 441 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]': finished 2026-03-10T10:17:00.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:00 vm03 bash[21274]: audit 2026-03-10T10:16:59.709063+0000 mon.a (mon.0) 441 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]': finished 2026-03-10T10:17:00.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:00 vm03 bash[21274]: cluster 2026-03-10T10:16:59.712306+0000 mon.a (mon.0) 442 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-10T10:17:00.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:00 vm03 bash[21274]: cluster 2026-03-10T10:16:59.712306+0000 mon.a (mon.0) 442 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-10T10:17:00.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:00 vm03 bash[21274]: audit 2026-03-10T10:16:59.712437+0000 mon.a (mon.0) 443 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:00.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:00 vm03 bash[21274]: audit 2026-03-10T10:16:59.712437+0000 mon.a (mon.0) 443 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:00.417 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:00 vm00 bash[20709]: audit 2026-03-10T10:16:59.706040+0000 mon.a (mon.0) 440 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]: dispatch 2026-03-10T10:17:00.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:00 vm00 bash[20709]: audit 2026-03-10T10:16:59.706040+0000 mon.a (mon.0) 440 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]: dispatch 2026-03-10T10:17:00.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:00 vm00 bash[20709]: audit 2026-03-10T10:16:59.706453+0000 mon.b (mon.2) 3 : audit [INF] from='client.? 192.168.123.103:0/1012162837' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]: dispatch 2026-03-10T10:17:00.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:00 vm00 bash[20709]: audit 2026-03-10T10:16:59.706453+0000 mon.b (mon.2) 3 : audit [INF] from='client.? 192.168.123.103:0/1012162837' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]: dispatch 2026-03-10T10:17:00.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:00 vm00 bash[20709]: audit 2026-03-10T10:16:59.709063+0000 mon.a (mon.0) 441 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]': finished 2026-03-10T10:17:00.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:00 vm00 bash[20709]: audit 2026-03-10T10:16:59.709063+0000 mon.a (mon.0) 441 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]': finished 2026-03-10T10:17:00.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:00 vm00 bash[20709]: cluster 2026-03-10T10:16:59.712306+0000 mon.a (mon.0) 442 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-10T10:17:00.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:00 vm00 bash[20709]: cluster 2026-03-10T10:16:59.712306+0000 mon.a (mon.0) 442 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-10T10:17:00.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:00 vm00 bash[20709]: audit 2026-03-10T10:16:59.712437+0000 mon.a (mon.0) 443 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:00.418 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:00 vm00 bash[20709]: audit 2026-03-10T10:16:59.712437+0000 mon.a (mon.0) 443 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:00.505 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:00 vm09 bash[20916]: audit 2026-03-10T10:16:59.706040+0000 mon.a (mon.0) 440 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]: dispatch 2026-03-10T10:17:00.505 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:00 vm09 bash[20916]: audit 2026-03-10T10:16:59.706040+0000 mon.a (mon.0) 440 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]: dispatch 2026-03-10T10:17:00.505 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:00 vm09 bash[20916]: audit 2026-03-10T10:16:59.706453+0000 mon.b (mon.2) 3 : audit [INF] from='client.? 192.168.123.103:0/1012162837' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]: dispatch 2026-03-10T10:17:00.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:00 vm09 bash[20916]: audit 2026-03-10T10:16:59.706453+0000 mon.b (mon.2) 3 : audit [INF] from='client.? 192.168.123.103:0/1012162837' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]: dispatch 2026-03-10T10:17:00.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:00 vm09 bash[20916]: audit 2026-03-10T10:16:59.709063+0000 mon.a (mon.0) 441 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]': finished 2026-03-10T10:17:00.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:00 vm09 bash[20916]: audit 2026-03-10T10:16:59.709063+0000 mon.a (mon.0) 441 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "48ce338d-486c-428f-88f7-6f089e65a594"}]': finished 2026-03-10T10:17:00.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:00 vm09 bash[20916]: cluster 2026-03-10T10:16:59.712306+0000 mon.a (mon.0) 442 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-10T10:17:00.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:00 vm09 bash[20916]: cluster 2026-03-10T10:16:59.712306+0000 mon.a (mon.0) 442 : cluster [DBG] osdmap e10: 2 total, 1 up, 2 in 2026-03-10T10:17:00.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:00 vm09 bash[20916]: audit 2026-03-10T10:16:59.712437+0000 mon.a (mon.0) 443 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:00.506 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:00 vm09 bash[20916]: audit 2026-03-10T10:16:59.712437+0000 mon.a (mon.0) 443 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:01.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:01 vm03 bash[21274]: cluster 2026-03-10T10:16:59.773294+0000 mgr.a (mgr.14150) 120 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:01.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:01 vm03 bash[21274]: cluster 2026-03-10T10:16:59.773294+0000 mgr.a (mgr.14150) 120 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:01.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:01 vm03 bash[21274]: audit 2026-03-10T10:17:00.323604+0000 mon.b (mon.2) 4 : audit [DBG] from='client.? 192.168.123.103:0/2915085576' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:17:01.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:01 vm03 bash[21274]: audit 2026-03-10T10:17:00.323604+0000 mon.b (mon.2) 4 : audit [DBG] from='client.? 192.168.123.103:0/2915085576' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:17:01.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:01 vm00 bash[20709]: cluster 2026-03-10T10:16:59.773294+0000 mgr.a (mgr.14150) 120 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:01.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:01 vm00 bash[20709]: cluster 2026-03-10T10:16:59.773294+0000 mgr.a (mgr.14150) 120 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:01.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:01 vm00 bash[20709]: audit 2026-03-10T10:17:00.323604+0000 mon.b (mon.2) 4 : audit [DBG] from='client.? 192.168.123.103:0/2915085576' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:17:01.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:01 vm00 bash[20709]: audit 2026-03-10T10:17:00.323604+0000 mon.b (mon.2) 4 : audit [DBG] from='client.? 192.168.123.103:0/2915085576' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:17:01.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:01 vm09 bash[20916]: cluster 2026-03-10T10:16:59.773294+0000 mgr.a (mgr.14150) 120 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:01.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:01 vm09 bash[20916]: cluster 2026-03-10T10:16:59.773294+0000 mgr.a (mgr.14150) 120 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:01.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:01 vm09 bash[20916]: audit 2026-03-10T10:17:00.323604+0000 mon.b (mon.2) 4 : audit [DBG] from='client.? 192.168.123.103:0/2915085576' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:17:01.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:01 vm09 bash[20916]: audit 2026-03-10T10:17:00.323604+0000 mon.b (mon.2) 4 : audit [DBG] from='client.? 192.168.123.103:0/2915085576' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:17:03.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:03 vm03 bash[21274]: cluster 2026-03-10T10:17:01.773515+0000 mgr.a (mgr.14150) 121 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:03.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:03 vm03 bash[21274]: cluster 2026-03-10T10:17:01.773515+0000 mgr.a (mgr.14150) 121 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:03.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:03 vm00 bash[20709]: cluster 2026-03-10T10:17:01.773515+0000 mgr.a (mgr.14150) 121 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:03.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:03 vm00 bash[20709]: cluster 2026-03-10T10:17:01.773515+0000 mgr.a (mgr.14150) 121 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:03.755 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:03 vm09 bash[20916]: cluster 2026-03-10T10:17:01.773515+0000 mgr.a (mgr.14150) 121 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:03.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:03 vm09 bash[20916]: cluster 2026-03-10T10:17:01.773515+0000 mgr.a (mgr.14150) 121 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:05.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:05 vm03 bash[21274]: cluster 2026-03-10T10:17:03.773794+0000 mgr.a (mgr.14150) 122 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:05.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:05 vm03 bash[21274]: cluster 2026-03-10T10:17:03.773794+0000 mgr.a (mgr.14150) 122 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:05.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:05 vm00 bash[20709]: cluster 2026-03-10T10:17:03.773794+0000 mgr.a (mgr.14150) 122 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:05.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:05 vm00 bash[20709]: cluster 2026-03-10T10:17:03.773794+0000 mgr.a (mgr.14150) 122 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:05.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:05 vm09 bash[20916]: cluster 2026-03-10T10:17:03.773794+0000 mgr.a (mgr.14150) 122 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:05.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:05 vm09 bash[20916]: cluster 2026-03-10T10:17:03.773794+0000 mgr.a (mgr.14150) 122 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:07.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:07 vm03 bash[21274]: cluster 2026-03-10T10:17:05.774073+0000 mgr.a (mgr.14150) 123 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:07.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:07 vm03 bash[21274]: cluster 2026-03-10T10:17:05.774073+0000 mgr.a (mgr.14150) 123 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:07.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:07 vm00 bash[20709]: cluster 2026-03-10T10:17:05.774073+0000 mgr.a (mgr.14150) 123 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:07.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:07 vm00 bash[20709]: cluster 2026-03-10T10:17:05.774073+0000 mgr.a (mgr.14150) 123 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:07.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:07 vm09 bash[20916]: cluster 2026-03-10T10:17:05.774073+0000 mgr.a (mgr.14150) 123 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:07.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:07 vm09 bash[20916]: cluster 2026-03-10T10:17:05.774073+0000 mgr.a (mgr.14150) 123 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:09.406 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:09 vm03 bash[21274]: cluster 2026-03-10T10:17:07.774328+0000 mgr.a (mgr.14150) 124 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:09.406 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:09 vm03 bash[21274]: cluster 2026-03-10T10:17:07.774328+0000 mgr.a (mgr.14150) 124 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:09.406 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:09 vm03 bash[21274]: audit 2026-03-10T10:17:08.574723+0000 mon.a (mon.0) 444 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T10:17:09.406 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:09 vm03 bash[21274]: audit 2026-03-10T10:17:08.574723+0000 mon.a (mon.0) 444 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T10:17:09.406 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:09 vm03 bash[21274]: audit 2026-03-10T10:17:08.575253+0000 mon.a (mon.0) 445 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:09.406 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:09 vm03 bash[21274]: audit 2026-03-10T10:17:08.575253+0000 mon.a (mon.0) 445 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:09.406 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:09 vm03 bash[21274]: cephadm 2026-03-10T10:17:08.575667+0000 mgr.a (mgr.14150) 125 : cephadm [INF] Deploying daemon osd.1 on vm03 2026-03-10T10:17:09.407 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:09 vm03 bash[21274]: cephadm 2026-03-10T10:17:08.575667+0000 mgr.a (mgr.14150) 125 : cephadm [INF] Deploying daemon osd.1 on vm03 2026-03-10T10:17:09.407 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:09 vm03 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:17:09.407 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:17:09 vm03 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:17:09.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:09 vm00 bash[20709]: cluster 2026-03-10T10:17:07.774328+0000 mgr.a (mgr.14150) 124 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:09.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:09 vm00 bash[20709]: cluster 2026-03-10T10:17:07.774328+0000 mgr.a (mgr.14150) 124 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:09.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:09 vm00 bash[20709]: audit 2026-03-10T10:17:08.574723+0000 mon.a (mon.0) 444 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T10:17:09.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:09 vm00 bash[20709]: audit 2026-03-10T10:17:08.574723+0000 mon.a (mon.0) 444 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T10:17:09.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:09 vm00 bash[20709]: audit 2026-03-10T10:17:08.575253+0000 mon.a (mon.0) 445 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:09.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:09 vm00 bash[20709]: audit 2026-03-10T10:17:08.575253+0000 mon.a (mon.0) 445 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:09.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:09 vm00 bash[20709]: cephadm 2026-03-10T10:17:08.575667+0000 mgr.a (mgr.14150) 125 : cephadm [INF] Deploying daemon osd.1 on vm03 2026-03-10T10:17:09.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:09 vm00 bash[20709]: cephadm 2026-03-10T10:17:08.575667+0000 mgr.a (mgr.14150) 125 : cephadm [INF] Deploying daemon osd.1 on vm03 2026-03-10T10:17:09.696 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:09 vm03 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:17:09.696 INFO:journalctl@ceph.mgr.b.vm03.stdout:Mar 10 10:17:09 vm03 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:17:09.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:09 vm09 bash[20916]: cluster 2026-03-10T10:17:07.774328+0000 mgr.a (mgr.14150) 124 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:09.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:09 vm09 bash[20916]: cluster 2026-03-10T10:17:07.774328+0000 mgr.a (mgr.14150) 124 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:09.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:09 vm09 bash[20916]: audit 2026-03-10T10:17:08.574723+0000 mon.a (mon.0) 444 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T10:17:09.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:09 vm09 bash[20916]: audit 2026-03-10T10:17:08.574723+0000 mon.a (mon.0) 444 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T10:17:09.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:09 vm09 bash[20916]: audit 2026-03-10T10:17:08.575253+0000 mon.a (mon.0) 445 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:09.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:09 vm09 bash[20916]: audit 2026-03-10T10:17:08.575253+0000 mon.a (mon.0) 445 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:09.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:09 vm09 bash[20916]: cephadm 2026-03-10T10:17:08.575667+0000 mgr.a (mgr.14150) 125 : cephadm [INF] Deploying daemon osd.1 on vm03 2026-03-10T10:17:09.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:09 vm09 bash[20916]: cephadm 2026-03-10T10:17:08.575667+0000 mgr.a (mgr.14150) 125 : cephadm [INF] Deploying daemon osd.1 on vm03 2026-03-10T10:17:10.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:10 vm03 bash[21274]: audit 2026-03-10T10:17:09.634989+0000 mon.a (mon.0) 446 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:10.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:10 vm03 bash[21274]: audit 2026-03-10T10:17:09.634989+0000 mon.a (mon.0) 446 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:10.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:10 vm03 bash[21274]: audit 2026-03-10T10:17:09.637274+0000 mon.a (mon.0) 447 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:10.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:10 vm03 bash[21274]: audit 2026-03-10T10:17:09.637274+0000 mon.a (mon.0) 447 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:10.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:10 vm03 bash[21274]: audit 2026-03-10T10:17:09.637812+0000 mon.a (mon.0) 448 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:10.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:10 vm03 bash[21274]: audit 2026-03-10T10:17:09.637812+0000 mon.a (mon.0) 448 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:10.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:10 vm03 bash[21274]: audit 2026-03-10T10:17:09.641375+0000 mon.a (mon.0) 449 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:10.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:10 vm03 bash[21274]: audit 2026-03-10T10:17:09.641375+0000 mon.a (mon.0) 449 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:10.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:10 vm03 bash[21274]: audit 2026-03-10T10:17:09.645337+0000 mon.a (mon.0) 450 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:10.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:10 vm03 bash[21274]: audit 2026-03-10T10:17:09.645337+0000 mon.a (mon.0) 450 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:10.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:10 vm03 bash[21274]: audit 2026-03-10T10:17:09.650767+0000 mon.a (mon.0) 451 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:10.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:10 vm03 bash[21274]: audit 2026-03-10T10:17:09.650767+0000 mon.a (mon.0) 451 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:10.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:10 vm00 bash[20709]: audit 2026-03-10T10:17:09.634989+0000 mon.a (mon.0) 446 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:10.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:10 vm00 bash[20709]: audit 2026-03-10T10:17:09.634989+0000 mon.a (mon.0) 446 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:10.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:10 vm00 bash[20709]: audit 2026-03-10T10:17:09.637274+0000 mon.a (mon.0) 447 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:10.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:10 vm00 bash[20709]: audit 2026-03-10T10:17:09.637274+0000 mon.a (mon.0) 447 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:10.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:10 vm00 bash[20709]: audit 2026-03-10T10:17:09.637812+0000 mon.a (mon.0) 448 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:10.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:10 vm00 bash[20709]: audit 2026-03-10T10:17:09.637812+0000 mon.a (mon.0) 448 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:10.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:10 vm00 bash[20709]: audit 2026-03-10T10:17:09.641375+0000 mon.a (mon.0) 449 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:10.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:10 vm00 bash[20709]: audit 2026-03-10T10:17:09.641375+0000 mon.a (mon.0) 449 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:10.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:10 vm00 bash[20709]: audit 2026-03-10T10:17:09.645337+0000 mon.a (mon.0) 450 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:10.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:10 vm00 bash[20709]: audit 2026-03-10T10:17:09.645337+0000 mon.a (mon.0) 450 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:10.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:10 vm00 bash[20709]: audit 2026-03-10T10:17:09.650767+0000 mon.a (mon.0) 451 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:10.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:10 vm00 bash[20709]: audit 2026-03-10T10:17:09.650767+0000 mon.a (mon.0) 451 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:10.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:10 vm09 bash[20916]: audit 2026-03-10T10:17:09.634989+0000 mon.a (mon.0) 446 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:10.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:10 vm09 bash[20916]: audit 2026-03-10T10:17:09.634989+0000 mon.a (mon.0) 446 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:10.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:10 vm09 bash[20916]: audit 2026-03-10T10:17:09.637274+0000 mon.a (mon.0) 447 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:10.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:10 vm09 bash[20916]: audit 2026-03-10T10:17:09.637274+0000 mon.a (mon.0) 447 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:10.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:10 vm09 bash[20916]: audit 2026-03-10T10:17:09.637812+0000 mon.a (mon.0) 448 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:10.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:10 vm09 bash[20916]: audit 2026-03-10T10:17:09.637812+0000 mon.a (mon.0) 448 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:10.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:10 vm09 bash[20916]: audit 2026-03-10T10:17:09.641375+0000 mon.a (mon.0) 449 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:10.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:10 vm09 bash[20916]: audit 2026-03-10T10:17:09.641375+0000 mon.a (mon.0) 449 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:10.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:10 vm09 bash[20916]: audit 2026-03-10T10:17:09.645337+0000 mon.a (mon.0) 450 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:10.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:10 vm09 bash[20916]: audit 2026-03-10T10:17:09.645337+0000 mon.a (mon.0) 450 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:10.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:10 vm09 bash[20916]: audit 2026-03-10T10:17:09.650767+0000 mon.a (mon.0) 451 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:10.756 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:10 vm09 bash[20916]: audit 2026-03-10T10:17:09.650767+0000 mon.a (mon.0) 451 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:11.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:11 vm00 bash[20709]: cluster 2026-03-10T10:17:09.774580+0000 mgr.a (mgr.14150) 126 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:11.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:11 vm00 bash[20709]: cluster 2026-03-10T10:17:09.774580+0000 mgr.a (mgr.14150) 126 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:12.004 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:11 vm03 bash[21274]: cluster 2026-03-10T10:17:09.774580+0000 mgr.a (mgr.14150) 126 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:12.004 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:11 vm03 bash[21274]: cluster 2026-03-10T10:17:09.774580+0000 mgr.a (mgr.14150) 126 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:12.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:11 vm09 bash[20916]: cluster 2026-03-10T10:17:09.774580+0000 mgr.a (mgr.14150) 126 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:12.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:11 vm09 bash[20916]: cluster 2026-03-10T10:17:09.774580+0000 mgr.a (mgr.14150) 126 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:13.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:13 vm03 bash[21274]: cluster 2026-03-10T10:17:11.774772+0000 mgr.a (mgr.14150) 127 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:13.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:13 vm03 bash[21274]: cluster 2026-03-10T10:17:11.774772+0000 mgr.a (mgr.14150) 127 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:13.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:13 vm03 bash[21274]: audit 2026-03-10T10:17:13.114662+0000 mon.a (mon.0) 452 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:17:13.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:13 vm03 bash[21274]: audit 2026-03-10T10:17:13.114662+0000 mon.a (mon.0) 452 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:17:13.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:13 vm03 bash[21274]: audit 2026-03-10T10:17:13.115878+0000 mon.c (mon.1) 7 : audit [INF] from='osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:17:13.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:13 vm03 bash[21274]: audit 2026-03-10T10:17:13.115878+0000 mon.c (mon.1) 7 : audit [INF] from='osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:17:13.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:13 vm00 bash[20709]: cluster 2026-03-10T10:17:11.774772+0000 mgr.a (mgr.14150) 127 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:13.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:13 vm00 bash[20709]: cluster 2026-03-10T10:17:11.774772+0000 mgr.a (mgr.14150) 127 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:13.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:13 vm00 bash[20709]: audit 2026-03-10T10:17:13.114662+0000 mon.a (mon.0) 452 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:17:13.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:13 vm00 bash[20709]: audit 2026-03-10T10:17:13.114662+0000 mon.a (mon.0) 452 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:17:13.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:13 vm00 bash[20709]: audit 2026-03-10T10:17:13.115878+0000 mon.c (mon.1) 7 : audit [INF] from='osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:17:13.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:13 vm00 bash[20709]: audit 2026-03-10T10:17:13.115878+0000 mon.c (mon.1) 7 : audit [INF] from='osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:17:14.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:13 vm09 bash[20916]: cluster 2026-03-10T10:17:11.774772+0000 mgr.a (mgr.14150) 127 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:14.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:13 vm09 bash[20916]: cluster 2026-03-10T10:17:11.774772+0000 mgr.a (mgr.14150) 127 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:14.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:13 vm09 bash[20916]: audit 2026-03-10T10:17:13.114662+0000 mon.a (mon.0) 452 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:17:14.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:13 vm09 bash[20916]: audit 2026-03-10T10:17:13.114662+0000 mon.a (mon.0) 452 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:17:14.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:13 vm09 bash[20916]: audit 2026-03-10T10:17:13.115878+0000 mon.c (mon.1) 7 : audit [INF] from='osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:17:14.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:13 vm09 bash[20916]: audit 2026-03-10T10:17:13.115878+0000 mon.c (mon.1) 7 : audit [INF] from='osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:17:15.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:14 vm09 bash[20916]: audit 2026-03-10T10:17:13.660672+0000 mon.a (mon.0) 453 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T10:17:15.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:14 vm09 bash[20916]: audit 2026-03-10T10:17:13.660672+0000 mon.a (mon.0) 453 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T10:17:15.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:14 vm09 bash[20916]: cluster 2026-03-10T10:17:13.663294+0000 mon.a (mon.0) 454 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-10T10:17:15.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:14 vm09 bash[20916]: cluster 2026-03-10T10:17:13.663294+0000 mon.a (mon.0) 454 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-10T10:17:15.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:14 vm09 bash[20916]: audit 2026-03-10T10:17:13.664091+0000 mon.a (mon.0) 455 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:15.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:14 vm09 bash[20916]: audit 2026-03-10T10:17:13.664091+0000 mon.a (mon.0) 455 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:15.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:14 vm09 bash[20916]: audit 2026-03-10T10:17:13.666306+0000 mon.c (mon.1) 8 : audit [INF] from='osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-10T10:17:15.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:14 vm09 bash[20916]: audit 2026-03-10T10:17:13.666306+0000 mon.c (mon.1) 8 : audit [INF] from='osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-10T10:17:15.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:14 vm09 bash[20916]: audit 2026-03-10T10:17:13.669294+0000 mon.a (mon.0) 456 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-10T10:17:15.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:14 vm09 bash[20916]: audit 2026-03-10T10:17:13.669294+0000 mon.a (mon.0) 456 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-10T10:17:15.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:14 vm03 bash[21274]: audit 2026-03-10T10:17:13.660672+0000 mon.a (mon.0) 453 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T10:17:15.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:14 vm03 bash[21274]: audit 2026-03-10T10:17:13.660672+0000 mon.a (mon.0) 453 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T10:17:15.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:14 vm03 bash[21274]: cluster 2026-03-10T10:17:13.663294+0000 mon.a (mon.0) 454 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-10T10:17:15.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:14 vm03 bash[21274]: cluster 2026-03-10T10:17:13.663294+0000 mon.a (mon.0) 454 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-10T10:17:15.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:14 vm03 bash[21274]: audit 2026-03-10T10:17:13.664091+0000 mon.a (mon.0) 455 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:15.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:14 vm03 bash[21274]: audit 2026-03-10T10:17:13.664091+0000 mon.a (mon.0) 455 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:15.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:14 vm03 bash[21274]: audit 2026-03-10T10:17:13.666306+0000 mon.c (mon.1) 8 : audit [INF] from='osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-10T10:17:15.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:14 vm03 bash[21274]: audit 2026-03-10T10:17:13.666306+0000 mon.c (mon.1) 8 : audit [INF] from='osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-10T10:17:15.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:14 vm03 bash[21274]: audit 2026-03-10T10:17:13.669294+0000 mon.a (mon.0) 456 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-10T10:17:15.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:14 vm03 bash[21274]: audit 2026-03-10T10:17:13.669294+0000 mon.a (mon.0) 456 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-10T10:17:15.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:14 vm00 bash[20709]: audit 2026-03-10T10:17:13.660672+0000 mon.a (mon.0) 453 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T10:17:15.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:14 vm00 bash[20709]: audit 2026-03-10T10:17:13.660672+0000 mon.a (mon.0) 453 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T10:17:15.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:14 vm00 bash[20709]: cluster 2026-03-10T10:17:13.663294+0000 mon.a (mon.0) 454 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-10T10:17:15.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:14 vm00 bash[20709]: cluster 2026-03-10T10:17:13.663294+0000 mon.a (mon.0) 454 : cluster [DBG] osdmap e11: 2 total, 1 up, 2 in 2026-03-10T10:17:15.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:14 vm00 bash[20709]: audit 2026-03-10T10:17:13.664091+0000 mon.a (mon.0) 455 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:15.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:14 vm00 bash[20709]: audit 2026-03-10T10:17:13.664091+0000 mon.a (mon.0) 455 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:15.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:14 vm00 bash[20709]: audit 2026-03-10T10:17:13.666306+0000 mon.c (mon.1) 8 : audit [INF] from='osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-10T10:17:15.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:14 vm00 bash[20709]: audit 2026-03-10T10:17:13.666306+0000 mon.c (mon.1) 8 : audit [INF] from='osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-10T10:17:15.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:14 vm00 bash[20709]: audit 2026-03-10T10:17:13.669294+0000 mon.a (mon.0) 456 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-10T10:17:15.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:14 vm00 bash[20709]: audit 2026-03-10T10:17:13.669294+0000 mon.a (mon.0) 456 : audit [INF] from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-10T10:17:15.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:15 vm03 bash[21274]: cluster 2026-03-10T10:17:13.775057+0000 mgr.a (mgr.14150) 128 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:15.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:15 vm03 bash[21274]: cluster 2026-03-10T10:17:13.775057+0000 mgr.a (mgr.14150) 128 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:15.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:15 vm03 bash[21274]: audit 2026-03-10T10:17:14.669038+0000 mon.a (mon.0) 457 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-10T10:17:15.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:15 vm03 bash[21274]: audit 2026-03-10T10:17:14.669038+0000 mon.a (mon.0) 457 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-10T10:17:15.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:15 vm03 bash[21274]: cluster 2026-03-10T10:17:14.670807+0000 mon.a (mon.0) 458 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-10T10:17:15.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:15 vm03 bash[21274]: cluster 2026-03-10T10:17:14.670807+0000 mon.a (mon.0) 458 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-10T10:17:15.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:15 vm03 bash[21274]: audit 2026-03-10T10:17:14.671531+0000 mon.a (mon.0) 459 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:15.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:15 vm03 bash[21274]: audit 2026-03-10T10:17:14.671531+0000 mon.a (mon.0) 459 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:15.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:15 vm03 bash[21274]: audit 2026-03-10T10:17:14.678647+0000 mon.a (mon.0) 460 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:15.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:15 vm03 bash[21274]: audit 2026-03-10T10:17:14.678647+0000 mon.a (mon.0) 460 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:16.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:15 vm09 bash[20916]: cluster 2026-03-10T10:17:13.775057+0000 mgr.a (mgr.14150) 128 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:16.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:15 vm09 bash[20916]: cluster 2026-03-10T10:17:13.775057+0000 mgr.a (mgr.14150) 128 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:16.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:15 vm09 bash[20916]: audit 2026-03-10T10:17:14.669038+0000 mon.a (mon.0) 457 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-10T10:17:16.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:15 vm09 bash[20916]: audit 2026-03-10T10:17:14.669038+0000 mon.a (mon.0) 457 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-10T10:17:16.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:15 vm09 bash[20916]: cluster 2026-03-10T10:17:14.670807+0000 mon.a (mon.0) 458 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-10T10:17:16.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:15 vm09 bash[20916]: cluster 2026-03-10T10:17:14.670807+0000 mon.a (mon.0) 458 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-10T10:17:16.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:15 vm09 bash[20916]: audit 2026-03-10T10:17:14.671531+0000 mon.a (mon.0) 459 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:16.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:15 vm09 bash[20916]: audit 2026-03-10T10:17:14.671531+0000 mon.a (mon.0) 459 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:16.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:15 vm09 bash[20916]: audit 2026-03-10T10:17:14.678647+0000 mon.a (mon.0) 460 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:16.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:15 vm09 bash[20916]: audit 2026-03-10T10:17:14.678647+0000 mon.a (mon.0) 460 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:16.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:15 vm00 bash[20709]: cluster 2026-03-10T10:17:13.775057+0000 mgr.a (mgr.14150) 128 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:16.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:15 vm00 bash[20709]: cluster 2026-03-10T10:17:13.775057+0000 mgr.a (mgr.14150) 128 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:16.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:15 vm00 bash[20709]: audit 2026-03-10T10:17:14.669038+0000 mon.a (mon.0) 457 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-10T10:17:16.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:15 vm00 bash[20709]: audit 2026-03-10T10:17:14.669038+0000 mon.a (mon.0) 457 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-10T10:17:16.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:15 vm00 bash[20709]: cluster 2026-03-10T10:17:14.670807+0000 mon.a (mon.0) 458 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-10T10:17:16.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:15 vm00 bash[20709]: cluster 2026-03-10T10:17:14.670807+0000 mon.a (mon.0) 458 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-10T10:17:16.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:15 vm00 bash[20709]: audit 2026-03-10T10:17:14.671531+0000 mon.a (mon.0) 459 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:16.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:15 vm00 bash[20709]: audit 2026-03-10T10:17:14.671531+0000 mon.a (mon.0) 459 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:16.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:15 vm00 bash[20709]: audit 2026-03-10T10:17:14.678647+0000 mon.a (mon.0) 460 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:16.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:15 vm00 bash[20709]: audit 2026-03-10T10:17:14.678647+0000 mon.a (mon.0) 460 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:16.621 INFO:teuthology.orchestra.run.vm03.stdout:Created osd(s) 1 on host 'vm03' 2026-03-10T10:17:16.676 DEBUG:teuthology.orchestra.run.vm03:osd.1> sudo journalctl -f -n 0 -u ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@osd.1.service 2026-03-10T10:17:16.677 INFO:tasks.cephadm:Deploying osd.2 on vm09 with /dev/vde... 2026-03-10T10:17:16.677 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- lvm zap /dev/vde 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: cluster 2026-03-10T10:17:14.101240+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: cluster 2026-03-10T10:17:14.101240+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: cluster 2026-03-10T10:17:14.101289+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: cluster 2026-03-10T10:17:14.101289+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:15.677727+0000 mon.a (mon.0) 461 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:15.677727+0000 mon.a (mon.0) 461 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: cluster 2026-03-10T10:17:15.680907+0000 mon.a (mon.0) 462 : cluster [INF] osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220] boot 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: cluster 2026-03-10T10:17:15.680907+0000 mon.a (mon.0) 462 : cluster [INF] osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220] boot 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: cluster 2026-03-10T10:17:15.681006+0000 mon.a (mon.0) 463 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: cluster 2026-03-10T10:17:15.681006+0000 mon.a (mon.0) 463 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:15.682171+0000 mon.a (mon.0) 464 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:15.682171+0000 mon.a (mon.0) 464 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.415566+0000 mon.a (mon.0) 465 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.415566+0000 mon.a (mon.0) 465 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.485184+0000 mon.a (mon.0) 466 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.485184+0000 mon.a (mon.0) 466 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.486771+0000 mon.a (mon.0) 467 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.486771+0000 mon.a (mon.0) 467 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.489426+0000 mon.a (mon.0) 468 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.489426+0000 mon.a (mon.0) 468 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.489953+0000 mon.a (mon.0) 469 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.489953+0000 mon.a (mon.0) 469 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.492345+0000 mon.a (mon.0) 470 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.492345+0000 mon.a (mon.0) 470 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.501131+0000 mon.a (mon.0) 471 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.501131+0000 mon.a (mon.0) 471 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.609605+0000 mon.a (mon.0) 472 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.609605+0000 mon.a (mon.0) 472 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.611492+0000 mon.a (mon.0) 473 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.611492+0000 mon.a (mon.0) 473 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.615007+0000 mon.a (mon.0) 474 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.615007+0000 mon.a (mon.0) 474 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.616811+0000 mon.a (mon.0) 475 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.616811+0000 mon.a (mon.0) 475 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.617221+0000 mon.a (mon.0) 476 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.617221+0000 mon.a (mon.0) 476 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.619060+0000 mon.a (mon.0) 477 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.619060+0000 mon.a (mon.0) 477 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.624863+0000 mon.a (mon.0) 478 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:16.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:16 vm03 bash[21274]: audit 2026-03-10T10:17:16.624863+0000 mon.a (mon.0) 478 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: cluster 2026-03-10T10:17:14.101240+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: cluster 2026-03-10T10:17:14.101240+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: cluster 2026-03-10T10:17:14.101289+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: cluster 2026-03-10T10:17:14.101289+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:15.677727+0000 mon.a (mon.0) 461 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:15.677727+0000 mon.a (mon.0) 461 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: cluster 2026-03-10T10:17:15.680907+0000 mon.a (mon.0) 462 : cluster [INF] osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220] boot 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: cluster 2026-03-10T10:17:15.680907+0000 mon.a (mon.0) 462 : cluster [INF] osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220] boot 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: cluster 2026-03-10T10:17:15.681006+0000 mon.a (mon.0) 463 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: cluster 2026-03-10T10:17:15.681006+0000 mon.a (mon.0) 463 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:15.682171+0000 mon.a (mon.0) 464 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:15.682171+0000 mon.a (mon.0) 464 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.415566+0000 mon.a (mon.0) 465 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.415566+0000 mon.a (mon.0) 465 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.485184+0000 mon.a (mon.0) 466 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.485184+0000 mon.a (mon.0) 466 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.486771+0000 mon.a (mon.0) 467 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.486771+0000 mon.a (mon.0) 467 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.489426+0000 mon.a (mon.0) 468 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.489426+0000 mon.a (mon.0) 468 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.489953+0000 mon.a (mon.0) 469 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.489953+0000 mon.a (mon.0) 469 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.492345+0000 mon.a (mon.0) 470 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.492345+0000 mon.a (mon.0) 470 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.501131+0000 mon.a (mon.0) 471 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.501131+0000 mon.a (mon.0) 471 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.609605+0000 mon.a (mon.0) 472 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.609605+0000 mon.a (mon.0) 472 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.611492+0000 mon.a (mon.0) 473 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.611492+0000 mon.a (mon.0) 473 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.615007+0000 mon.a (mon.0) 474 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.615007+0000 mon.a (mon.0) 474 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.616811+0000 mon.a (mon.0) 475 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.616811+0000 mon.a (mon.0) 475 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.617221+0000 mon.a (mon.0) 476 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.617221+0000 mon.a (mon.0) 476 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.619060+0000 mon.a (mon.0) 477 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.619060+0000 mon.a (mon.0) 477 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.624863+0000 mon.a (mon.0) 478 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:16 vm09 bash[20916]: audit 2026-03-10T10:17:16.624863+0000 mon.a (mon.0) 478 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: cluster 2026-03-10T10:17:14.101240+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: cluster 2026-03-10T10:17:14.101240+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: cluster 2026-03-10T10:17:14.101289+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: cluster 2026-03-10T10:17:14.101289+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:15.677727+0000 mon.a (mon.0) 461 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:15.677727+0000 mon.a (mon.0) 461 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: cluster 2026-03-10T10:17:15.680907+0000 mon.a (mon.0) 462 : cluster [INF] osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220] boot 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: cluster 2026-03-10T10:17:15.680907+0000 mon.a (mon.0) 462 : cluster [INF] osd.1 [v2:192.168.123.103:6800/4141501220,v1:192.168.123.103:6801/4141501220] boot 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: cluster 2026-03-10T10:17:15.681006+0000 mon.a (mon.0) 463 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: cluster 2026-03-10T10:17:15.681006+0000 mon.a (mon.0) 463 : cluster [DBG] osdmap e13: 2 total, 2 up, 2 in 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:15.682171+0000 mon.a (mon.0) 464 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:15.682171+0000 mon.a (mon.0) 464 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.415566+0000 mon.a (mon.0) 465 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.415566+0000 mon.a (mon.0) 465 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.485184+0000 mon.a (mon.0) 466 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.485184+0000 mon.a (mon.0) 466 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.486771+0000 mon.a (mon.0) 467 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.486771+0000 mon.a (mon.0) 467 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.489426+0000 mon.a (mon.0) 468 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.489426+0000 mon.a (mon.0) 468 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.489953+0000 mon.a (mon.0) 469 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.489953+0000 mon.a (mon.0) 469 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.492345+0000 mon.a (mon.0) 470 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.492345+0000 mon.a (mon.0) 470 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.501131+0000 mon.a (mon.0) 471 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.501131+0000 mon.a (mon.0) 471 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.609605+0000 mon.a (mon.0) 472 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.609605+0000 mon.a (mon.0) 472 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.611492+0000 mon.a (mon.0) 473 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.611492+0000 mon.a (mon.0) 473 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.615007+0000 mon.a (mon.0) 474 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.615007+0000 mon.a (mon.0) 474 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.616811+0000 mon.a (mon.0) 475 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:17.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.616811+0000 mon.a (mon.0) 475 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:17.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.617221+0000 mon.a (mon.0) 476 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:17.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.617221+0000 mon.a (mon.0) 476 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:17.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.619060+0000 mon.a (mon.0) 477 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.619060+0000 mon.a (mon.0) 477 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.624863+0000 mon.a (mon.0) 478 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.168 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:16 vm00 bash[20709]: audit 2026-03-10T10:17:16.624863+0000 mon.a (mon.0) 478 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:17.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:17 vm03 bash[21274]: cluster 2026-03-10T10:17:15.775253+0000 mgr.a (mgr.14150) 129 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:17.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:17 vm03 bash[21274]: cluster 2026-03-10T10:17:15.775253+0000 mgr.a (mgr.14150) 129 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:17.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:17 vm03 bash[21274]: cephadm 2026-03-10T10:17:16.486108+0000 mgr.a (mgr.14150) 130 : cephadm [INF] Detected new or changed devices on vm03 2026-03-10T10:17:17.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:17 vm03 bash[21274]: cephadm 2026-03-10T10:17:16.486108+0000 mgr.a (mgr.14150) 130 : cephadm [INF] Detected new or changed devices on vm03 2026-03-10T10:17:17.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:17 vm03 bash[21274]: cluster 2026-03-10T10:17:16.693018+0000 mon.a (mon.0) 479 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T10:17:17.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:17 vm03 bash[21274]: cluster 2026-03-10T10:17:16.693018+0000 mon.a (mon.0) 479 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T10:17:18.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:17 vm09 bash[20916]: cluster 2026-03-10T10:17:15.775253+0000 mgr.a (mgr.14150) 129 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:18.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:17 vm09 bash[20916]: cluster 2026-03-10T10:17:15.775253+0000 mgr.a (mgr.14150) 129 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:18.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:17 vm09 bash[20916]: cephadm 2026-03-10T10:17:16.486108+0000 mgr.a (mgr.14150) 130 : cephadm [INF] Detected new or changed devices on vm03 2026-03-10T10:17:18.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:17 vm09 bash[20916]: cephadm 2026-03-10T10:17:16.486108+0000 mgr.a (mgr.14150) 130 : cephadm [INF] Detected new or changed devices on vm03 2026-03-10T10:17:18.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:17 vm09 bash[20916]: cluster 2026-03-10T10:17:16.693018+0000 mon.a (mon.0) 479 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T10:17:18.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:17 vm09 bash[20916]: cluster 2026-03-10T10:17:16.693018+0000 mon.a (mon.0) 479 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T10:17:18.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:17 vm00 bash[20709]: cluster 2026-03-10T10:17:15.775253+0000 mgr.a (mgr.14150) 129 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:18.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:17 vm00 bash[20709]: cluster 2026-03-10T10:17:15.775253+0000 mgr.a (mgr.14150) 129 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T10:17:18.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:17 vm00 bash[20709]: cephadm 2026-03-10T10:17:16.486108+0000 mgr.a (mgr.14150) 130 : cephadm [INF] Detected new or changed devices on vm03 2026-03-10T10:17:18.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:17 vm00 bash[20709]: cephadm 2026-03-10T10:17:16.486108+0000 mgr.a (mgr.14150) 130 : cephadm [INF] Detected new or changed devices on vm03 2026-03-10T10:17:18.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:17 vm00 bash[20709]: cluster 2026-03-10T10:17:16.693018+0000 mon.a (mon.0) 479 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T10:17:18.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:17 vm00 bash[20709]: cluster 2026-03-10T10:17:16.693018+0000 mon.a (mon.0) 479 : cluster [DBG] osdmap e14: 2 total, 2 up, 2 in 2026-03-10T10:17:19.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.693386+0000 mon.a (mon.0) 480 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.693386+0000 mon.a (mon.0) 480 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.704040+0000 mon.a (mon.0) 481 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.704040+0000 mon.a (mon.0) 481 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.707476+0000 mon.a (mon.0) 482 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.707476+0000 mon.a (mon.0) 482 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.711429+0000 mon.a (mon.0) 483 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.711429+0000 mon.a (mon.0) 483 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.712638+0000 mon.a (mon.0) 484 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:19.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.712638+0000 mon.a (mon.0) 484 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:19.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.713707+0000 mon.a (mon.0) 485 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:19.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.713707+0000 mon.a (mon.0) 485 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:19.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.714079+0000 mon.a (mon.0) 486 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:19.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.714079+0000 mon.a (mon.0) 486 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:19.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.715782+0000 mon.a (mon.0) 487 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.715782+0000 mon.a (mon.0) 487 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.719812+0000 mon.a (mon.0) 488 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.719812+0000 mon.a (mon.0) 488 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.723528+0000 mon.a (mon.0) 489 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:18 vm09 bash[20916]: audit 2026-03-10T10:17:17.723528+0000 mon.a (mon.0) 489 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.693386+0000 mon.a (mon.0) 480 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.693386+0000 mon.a (mon.0) 480 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.704040+0000 mon.a (mon.0) 481 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.704040+0000 mon.a (mon.0) 481 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.707476+0000 mon.a (mon.0) 482 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.707476+0000 mon.a (mon.0) 482 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.711429+0000 mon.a (mon.0) 483 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.711429+0000 mon.a (mon.0) 483 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.712638+0000 mon.a (mon.0) 484 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.712638+0000 mon.a (mon.0) 484 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.713707+0000 mon.a (mon.0) 485 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.713707+0000 mon.a (mon.0) 485 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.714079+0000 mon.a (mon.0) 486 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.714079+0000 mon.a (mon.0) 486 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.715782+0000 mon.a (mon.0) 487 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.715782+0000 mon.a (mon.0) 487 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.719812+0000 mon.a (mon.0) 488 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.719812+0000 mon.a (mon.0) 488 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.723528+0000 mon.a (mon.0) 489 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:18 vm03 bash[21274]: audit 2026-03-10T10:17:17.723528+0000 mon.a (mon.0) 489 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.693386+0000 mon.a (mon.0) 480 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.693386+0000 mon.a (mon.0) 480 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.704040+0000 mon.a (mon.0) 481 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.704040+0000 mon.a (mon.0) 481 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.707476+0000 mon.a (mon.0) 482 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.707476+0000 mon.a (mon.0) 482 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.711429+0000 mon.a (mon.0) 483 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.711429+0000 mon.a (mon.0) 483 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.712638+0000 mon.a (mon.0) 484 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.712638+0000 mon.a (mon.0) 484 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.713707+0000 mon.a (mon.0) 485 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.713707+0000 mon.a (mon.0) 485 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.714079+0000 mon.a (mon.0) 486 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.714079+0000 mon.a (mon.0) 486 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.715782+0000 mon.a (mon.0) 487 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.715782+0000 mon.a (mon.0) 487 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.719812+0000 mon.a (mon.0) 488 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.719812+0000 mon.a (mon.0) 488 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.723528+0000 mon.a (mon.0) 489 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:19.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:18 vm00 bash[20709]: audit 2026-03-10T10:17:17.723528+0000 mon.a (mon.0) 489 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:19 vm09 bash[20916]: cluster 2026-03-10T10:17:17.775469+0000 mgr.a (mgr.14150) 131 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:20.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:19 vm09 bash[20916]: cluster 2026-03-10T10:17:17.775469+0000 mgr.a (mgr.14150) 131 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:20.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:19 vm09 bash[20916]: audit 2026-03-10T10:17:19.169809+0000 mon.a (mon.0) 490 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:19 vm09 bash[20916]: audit 2026-03-10T10:17:19.169809+0000 mon.a (mon.0) 490 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:19 vm09 bash[20916]: audit 2026-03-10T10:17:19.172900+0000 mon.a (mon.0) 491 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:19 vm09 bash[20916]: audit 2026-03-10T10:17:19.172900+0000 mon.a (mon.0) 491 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:19 vm09 bash[20916]: audit 2026-03-10T10:17:19.176246+0000 mon.a (mon.0) 492 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:19 vm09 bash[20916]: audit 2026-03-10T10:17:19.176246+0000 mon.a (mon.0) 492 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:19 vm03 bash[21274]: cluster 2026-03-10T10:17:17.775469+0000 mgr.a (mgr.14150) 131 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:20.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:19 vm03 bash[21274]: cluster 2026-03-10T10:17:17.775469+0000 mgr.a (mgr.14150) 131 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:20.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:19 vm03 bash[21274]: audit 2026-03-10T10:17:19.169809+0000 mon.a (mon.0) 490 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:19 vm03 bash[21274]: audit 2026-03-10T10:17:19.169809+0000 mon.a (mon.0) 490 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:19 vm03 bash[21274]: audit 2026-03-10T10:17:19.172900+0000 mon.a (mon.0) 491 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:19 vm03 bash[21274]: audit 2026-03-10T10:17:19.172900+0000 mon.a (mon.0) 491 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:19 vm03 bash[21274]: audit 2026-03-10T10:17:19.176246+0000 mon.a (mon.0) 492 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:19 vm03 bash[21274]: audit 2026-03-10T10:17:19.176246+0000 mon.a (mon.0) 492 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:19 vm00 bash[20709]: cluster 2026-03-10T10:17:17.775469+0000 mgr.a (mgr.14150) 131 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:20.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:19 vm00 bash[20709]: cluster 2026-03-10T10:17:17.775469+0000 mgr.a (mgr.14150) 131 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:20.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:19 vm00 bash[20709]: audit 2026-03-10T10:17:19.169809+0000 mon.a (mon.0) 490 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:19 vm00 bash[20709]: audit 2026-03-10T10:17:19.169809+0000 mon.a (mon.0) 490 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:19 vm00 bash[20709]: audit 2026-03-10T10:17:19.172900+0000 mon.a (mon.0) 491 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:19 vm00 bash[20709]: audit 2026-03-10T10:17:19.172900+0000 mon.a (mon.0) 491 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:19 vm00 bash[20709]: audit 2026-03-10T10:17:19.176246+0000 mon.a (mon.0) 492 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:19 vm00 bash[20709]: audit 2026-03-10T10:17:19.176246+0000 mon.a (mon.0) 492 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:20.292 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.c/config 2026-03-10T10:17:21.110 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:17:21.125 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph orch daemon add osd vm09:/dev/vde 2026-03-10T10:17:22.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:21 vm09 bash[20916]: cluster 2026-03-10T10:17:19.775697+0000 mgr.a (mgr.14150) 132 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:22.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:21 vm09 bash[20916]: cluster 2026-03-10T10:17:19.775697+0000 mgr.a (mgr.14150) 132 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:22.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:21 vm03 bash[21274]: cluster 2026-03-10T10:17:19.775697+0000 mgr.a (mgr.14150) 132 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:22.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:21 vm03 bash[21274]: cluster 2026-03-10T10:17:19.775697+0000 mgr.a (mgr.14150) 132 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:22.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:21 vm00 bash[20709]: cluster 2026-03-10T10:17:19.775697+0000 mgr.a (mgr.14150) 132 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:22.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:21 vm00 bash[20709]: cluster 2026-03-10T10:17:19.775697+0000 mgr.a (mgr.14150) 132 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:24.006 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:23 vm09 bash[20916]: cluster 2026-03-10T10:17:21.775880+0000 mgr.a (mgr.14150) 133 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:24.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:23 vm09 bash[20916]: cluster 2026-03-10T10:17:21.775880+0000 mgr.a (mgr.14150) 133 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:24.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:23 vm03 bash[21274]: cluster 2026-03-10T10:17:21.775880+0000 mgr.a (mgr.14150) 133 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:24.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:23 vm03 bash[21274]: cluster 2026-03-10T10:17:21.775880+0000 mgr.a (mgr.14150) 133 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:24.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:23 vm00 bash[20709]: cluster 2026-03-10T10:17:21.775880+0000 mgr.a (mgr.14150) 133 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:24.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:23 vm00 bash[20709]: cluster 2026-03-10T10:17:21.775880+0000 mgr.a (mgr.14150) 133 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:24.731 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.c/config 2026-03-10T10:17:26.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:25 vm09 bash[20916]: cluster 2026-03-10T10:17:23.776075+0000 mgr.a (mgr.14150) 134 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:26.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:25 vm09 bash[20916]: cluster 2026-03-10T10:17:23.776075+0000 mgr.a (mgr.14150) 134 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:26.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:25 vm09 bash[20916]: audit 2026-03-10T10:17:25.148751+0000 mon.a (mon.0) 493 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:17:26.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:25 vm09 bash[20916]: audit 2026-03-10T10:17:25.148751+0000 mon.a (mon.0) 493 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:17:26.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:25 vm09 bash[20916]: audit 2026-03-10T10:17:25.149874+0000 mon.a (mon.0) 494 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:17:26.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:25 vm09 bash[20916]: audit 2026-03-10T10:17:25.149874+0000 mon.a (mon.0) 494 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:17:26.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:25 vm09 bash[20916]: audit 2026-03-10T10:17:25.150193+0000 mon.a (mon.0) 495 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:26.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:25 vm09 bash[20916]: audit 2026-03-10T10:17:25.150193+0000 mon.a (mon.0) 495 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:26.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:25 vm03 bash[21274]: cluster 2026-03-10T10:17:23.776075+0000 mgr.a (mgr.14150) 134 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:26.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:25 vm03 bash[21274]: cluster 2026-03-10T10:17:23.776075+0000 mgr.a (mgr.14150) 134 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:26.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:25 vm03 bash[21274]: audit 2026-03-10T10:17:25.148751+0000 mon.a (mon.0) 493 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:17:26.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:25 vm03 bash[21274]: audit 2026-03-10T10:17:25.148751+0000 mon.a (mon.0) 493 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:17:26.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:25 vm03 bash[21274]: audit 2026-03-10T10:17:25.149874+0000 mon.a (mon.0) 494 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:17:26.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:25 vm03 bash[21274]: audit 2026-03-10T10:17:25.149874+0000 mon.a (mon.0) 494 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:17:26.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:25 vm03 bash[21274]: audit 2026-03-10T10:17:25.150193+0000 mon.a (mon.0) 495 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:26.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:25 vm03 bash[21274]: audit 2026-03-10T10:17:25.150193+0000 mon.a (mon.0) 495 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:26.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:25 vm00 bash[20709]: cluster 2026-03-10T10:17:23.776075+0000 mgr.a (mgr.14150) 134 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:26.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:25 vm00 bash[20709]: cluster 2026-03-10T10:17:23.776075+0000 mgr.a (mgr.14150) 134 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:26.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:25 vm00 bash[20709]: audit 2026-03-10T10:17:25.148751+0000 mon.a (mon.0) 493 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:17:26.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:25 vm00 bash[20709]: audit 2026-03-10T10:17:25.148751+0000 mon.a (mon.0) 493 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:17:26.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:25 vm00 bash[20709]: audit 2026-03-10T10:17:25.149874+0000 mon.a (mon.0) 494 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:17:26.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:25 vm00 bash[20709]: audit 2026-03-10T10:17:25.149874+0000 mon.a (mon.0) 494 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:17:26.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:25 vm00 bash[20709]: audit 2026-03-10T10:17:25.150193+0000 mon.a (mon.0) 495 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:26.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:25 vm00 bash[20709]: audit 2026-03-10T10:17:25.150193+0000 mon.a (mon.0) 495 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:27.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:26 vm00 bash[20709]: audit 2026-03-10T10:17:25.147408+0000 mgr.a (mgr.14150) 135 : audit [DBG] from='client.24158 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:17:27.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:26 vm00 bash[20709]: audit 2026-03-10T10:17:25.147408+0000 mgr.a (mgr.14150) 135 : audit [DBG] from='client.24158 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:17:27.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:26 vm09 bash[20916]: audit 2026-03-10T10:17:25.147408+0000 mgr.a (mgr.14150) 135 : audit [DBG] from='client.24158 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:17:27.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:26 vm09 bash[20916]: audit 2026-03-10T10:17:25.147408+0000 mgr.a (mgr.14150) 135 : audit [DBG] from='client.24158 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:17:27.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:26 vm03 bash[21274]: audit 2026-03-10T10:17:25.147408+0000 mgr.a (mgr.14150) 135 : audit [DBG] from='client.24158 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:17:27.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:26 vm03 bash[21274]: audit 2026-03-10T10:17:25.147408+0000 mgr.a (mgr.14150) 135 : audit [DBG] from='client.24158 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:17:28.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:27 vm09 bash[20916]: cluster 2026-03-10T10:17:25.776259+0000 mgr.a (mgr.14150) 136 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:28.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:27 vm09 bash[20916]: cluster 2026-03-10T10:17:25.776259+0000 mgr.a (mgr.14150) 136 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:28.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:27 vm03 bash[21274]: cluster 2026-03-10T10:17:25.776259+0000 mgr.a (mgr.14150) 136 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:28.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:27 vm03 bash[21274]: cluster 2026-03-10T10:17:25.776259+0000 mgr.a (mgr.14150) 136 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:27 vm00 bash[20709]: cluster 2026-03-10T10:17:25.776259+0000 mgr.a (mgr.14150) 136 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:28.419 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:27 vm00 bash[20709]: cluster 2026-03-10T10:17:25.776259+0000 mgr.a (mgr.14150) 136 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:29.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:28 vm09 bash[20916]: cluster 2026-03-10T10:17:27.776476+0000 mgr.a (mgr.14150) 137 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:29.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:28 vm09 bash[20916]: cluster 2026-03-10T10:17:27.776476+0000 mgr.a (mgr.14150) 137 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:29.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:28 vm03 bash[21274]: cluster 2026-03-10T10:17:27.776476+0000 mgr.a (mgr.14150) 137 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:29.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:28 vm03 bash[21274]: cluster 2026-03-10T10:17:27.776476+0000 mgr.a (mgr.14150) 137 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:29.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:28 vm00 bash[20709]: cluster 2026-03-10T10:17:27.776476+0000 mgr.a (mgr.14150) 137 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:29.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:28 vm00 bash[20709]: cluster 2026-03-10T10:17:27.776476+0000 mgr.a (mgr.14150) 137 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:31.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:31 vm03 bash[21274]: cluster 2026-03-10T10:17:29.776763+0000 mgr.a (mgr.14150) 138 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:31.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:31 vm03 bash[21274]: cluster 2026-03-10T10:17:29.776763+0000 mgr.a (mgr.14150) 138 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:31.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:31 vm03 bash[21274]: audit 2026-03-10T10:17:30.591983+0000 mon.a (mon.0) 496 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]: dispatch 2026-03-10T10:17:31.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:31 vm03 bash[21274]: audit 2026-03-10T10:17:30.591983+0000 mon.a (mon.0) 496 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]: dispatch 2026-03-10T10:17:31.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:31 vm03 bash[21274]: audit 2026-03-10T10:17:30.592046+0000 mon.c (mon.1) 9 : audit [INF] from='client.? 192.168.123.109:0/1176550768' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]: dispatch 2026-03-10T10:17:31.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:31 vm03 bash[21274]: audit 2026-03-10T10:17:30.592046+0000 mon.c (mon.1) 9 : audit [INF] from='client.? 192.168.123.109:0/1176550768' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]: dispatch 2026-03-10T10:17:31.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:31 vm03 bash[21274]: audit 2026-03-10T10:17:30.594681+0000 mon.a (mon.0) 497 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]': finished 2026-03-10T10:17:31.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:31 vm03 bash[21274]: audit 2026-03-10T10:17:30.594681+0000 mon.a (mon.0) 497 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]': finished 2026-03-10T10:17:31.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:31 vm03 bash[21274]: cluster 2026-03-10T10:17:30.597882+0000 mon.a (mon.0) 498 : cluster [DBG] osdmap e15: 3 total, 2 up, 3 in 2026-03-10T10:17:31.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:31 vm03 bash[21274]: cluster 2026-03-10T10:17:30.597882+0000 mon.a (mon.0) 498 : cluster [DBG] osdmap e15: 3 total, 2 up, 3 in 2026-03-10T10:17:31.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:31 vm03 bash[21274]: audit 2026-03-10T10:17:30.598034+0000 mon.a (mon.0) 499 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:31.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:31 vm03 bash[21274]: audit 2026-03-10T10:17:30.598034+0000 mon.a (mon.0) 499 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:31.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:31 vm00 bash[20709]: cluster 2026-03-10T10:17:29.776763+0000 mgr.a (mgr.14150) 138 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:31.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:31 vm00 bash[20709]: cluster 2026-03-10T10:17:29.776763+0000 mgr.a (mgr.14150) 138 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:31.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:31 vm00 bash[20709]: audit 2026-03-10T10:17:30.591983+0000 mon.a (mon.0) 496 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]: dispatch 2026-03-10T10:17:31.417 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:31 vm00 bash[20709]: audit 2026-03-10T10:17:30.591983+0000 mon.a (mon.0) 496 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]: dispatch 2026-03-10T10:17:31.417 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:31 vm00 bash[20709]: audit 2026-03-10T10:17:30.592046+0000 mon.c (mon.1) 9 : audit [INF] from='client.? 192.168.123.109:0/1176550768' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]: dispatch 2026-03-10T10:17:31.417 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:31 vm00 bash[20709]: audit 2026-03-10T10:17:30.592046+0000 mon.c (mon.1) 9 : audit [INF] from='client.? 192.168.123.109:0/1176550768' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]: dispatch 2026-03-10T10:17:31.417 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:31 vm00 bash[20709]: audit 2026-03-10T10:17:30.594681+0000 mon.a (mon.0) 497 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]': finished 2026-03-10T10:17:31.417 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:31 vm00 bash[20709]: audit 2026-03-10T10:17:30.594681+0000 mon.a (mon.0) 497 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]': finished 2026-03-10T10:17:31.417 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:31 vm00 bash[20709]: cluster 2026-03-10T10:17:30.597882+0000 mon.a (mon.0) 498 : cluster [DBG] osdmap e15: 3 total, 2 up, 3 in 2026-03-10T10:17:31.417 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:31 vm00 bash[20709]: cluster 2026-03-10T10:17:30.597882+0000 mon.a (mon.0) 498 : cluster [DBG] osdmap e15: 3 total, 2 up, 3 in 2026-03-10T10:17:31.417 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:31 vm00 bash[20709]: audit 2026-03-10T10:17:30.598034+0000 mon.a (mon.0) 499 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:31.417 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:31 vm00 bash[20709]: audit 2026-03-10T10:17:30.598034+0000 mon.a (mon.0) 499 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:31.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:31 vm09 bash[20916]: cluster 2026-03-10T10:17:29.776763+0000 mgr.a (mgr.14150) 138 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:31.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:31 vm09 bash[20916]: cluster 2026-03-10T10:17:29.776763+0000 mgr.a (mgr.14150) 138 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:31.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:31 vm09 bash[20916]: audit 2026-03-10T10:17:30.591983+0000 mon.a (mon.0) 496 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]: dispatch 2026-03-10T10:17:31.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:31 vm09 bash[20916]: audit 2026-03-10T10:17:30.591983+0000 mon.a (mon.0) 496 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]: dispatch 2026-03-10T10:17:31.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:31 vm09 bash[20916]: audit 2026-03-10T10:17:30.592046+0000 mon.c (mon.1) 9 : audit [INF] from='client.? 192.168.123.109:0/1176550768' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]: dispatch 2026-03-10T10:17:31.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:31 vm09 bash[20916]: audit 2026-03-10T10:17:30.592046+0000 mon.c (mon.1) 9 : audit [INF] from='client.? 192.168.123.109:0/1176550768' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]: dispatch 2026-03-10T10:17:31.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:31 vm09 bash[20916]: audit 2026-03-10T10:17:30.594681+0000 mon.a (mon.0) 497 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]': finished 2026-03-10T10:17:31.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:31 vm09 bash[20916]: audit 2026-03-10T10:17:30.594681+0000 mon.a (mon.0) 497 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c1f82793-95a2-4beb-bb79-141603b483df"}]': finished 2026-03-10T10:17:31.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:31 vm09 bash[20916]: cluster 2026-03-10T10:17:30.597882+0000 mon.a (mon.0) 498 : cluster [DBG] osdmap e15: 3 total, 2 up, 3 in 2026-03-10T10:17:31.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:31 vm09 bash[20916]: cluster 2026-03-10T10:17:30.597882+0000 mon.a (mon.0) 498 : cluster [DBG] osdmap e15: 3 total, 2 up, 3 in 2026-03-10T10:17:31.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:31 vm09 bash[20916]: audit 2026-03-10T10:17:30.598034+0000 mon.a (mon.0) 499 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:31.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:31 vm09 bash[20916]: audit 2026-03-10T10:17:30.598034+0000 mon.a (mon.0) 499 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:32.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:32 vm03 bash[21274]: audit 2026-03-10T10:17:31.193053+0000 mon.b (mon.2) 5 : audit [DBG] from='client.? 192.168.123.109:0/1983654338' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:17:32.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:32 vm03 bash[21274]: audit 2026-03-10T10:17:31.193053+0000 mon.b (mon.2) 5 : audit [DBG] from='client.? 192.168.123.109:0/1983654338' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:17:32.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:32 vm00 bash[20709]: audit 2026-03-10T10:17:31.193053+0000 mon.b (mon.2) 5 : audit [DBG] from='client.? 192.168.123.109:0/1983654338' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:17:32.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:32 vm00 bash[20709]: audit 2026-03-10T10:17:31.193053+0000 mon.b (mon.2) 5 : audit [DBG] from='client.? 192.168.123.109:0/1983654338' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:17:32.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:32 vm09 bash[20916]: audit 2026-03-10T10:17:31.193053+0000 mon.b (mon.2) 5 : audit [DBG] from='client.? 192.168.123.109:0/1983654338' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:17:32.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:32 vm09 bash[20916]: audit 2026-03-10T10:17:31.193053+0000 mon.b (mon.2) 5 : audit [DBG] from='client.? 192.168.123.109:0/1983654338' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:17:33.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:33 vm03 bash[21274]: cluster 2026-03-10T10:17:31.776965+0000 mgr.a (mgr.14150) 139 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:33.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:33 vm03 bash[21274]: cluster 2026-03-10T10:17:31.776965+0000 mgr.a (mgr.14150) 139 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:33.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:33 vm00 bash[20709]: cluster 2026-03-10T10:17:31.776965+0000 mgr.a (mgr.14150) 139 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:33.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:33 vm00 bash[20709]: cluster 2026-03-10T10:17:31.776965+0000 mgr.a (mgr.14150) 139 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:33.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:33 vm09 bash[20916]: cluster 2026-03-10T10:17:31.776965+0000 mgr.a (mgr.14150) 139 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:33.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:33 vm09 bash[20916]: cluster 2026-03-10T10:17:31.776965+0000 mgr.a (mgr.14150) 139 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:35.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:35 vm03 bash[21274]: cluster 2026-03-10T10:17:33.777215+0000 mgr.a (mgr.14150) 140 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:35.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:35 vm03 bash[21274]: cluster 2026-03-10T10:17:33.777215+0000 mgr.a (mgr.14150) 140 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:35.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:35 vm00 bash[20709]: cluster 2026-03-10T10:17:33.777215+0000 mgr.a (mgr.14150) 140 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:35.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:35 vm00 bash[20709]: cluster 2026-03-10T10:17:33.777215+0000 mgr.a (mgr.14150) 140 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:35.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:35 vm09 bash[20916]: cluster 2026-03-10T10:17:33.777215+0000 mgr.a (mgr.14150) 140 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:35.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:35 vm09 bash[20916]: cluster 2026-03-10T10:17:33.777215+0000 mgr.a (mgr.14150) 140 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:37.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:37 vm03 bash[21274]: cluster 2026-03-10T10:17:35.777422+0000 mgr.a (mgr.14150) 141 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:37.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:37 vm03 bash[21274]: cluster 2026-03-10T10:17:35.777422+0000 mgr.a (mgr.14150) 141 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:37.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:37 vm00 bash[20709]: cluster 2026-03-10T10:17:35.777422+0000 mgr.a (mgr.14150) 141 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:37.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:37 vm00 bash[20709]: cluster 2026-03-10T10:17:35.777422+0000 mgr.a (mgr.14150) 141 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:37.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:37 vm09 bash[20916]: cluster 2026-03-10T10:17:35.777422+0000 mgr.a (mgr.14150) 141 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:37.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:37 vm09 bash[20916]: cluster 2026-03-10T10:17:35.777422+0000 mgr.a (mgr.14150) 141 : cluster [DBG] pgmap v80: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:38.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:38 vm03 bash[21274]: audit 2026-03-10T10:17:37.257930+0000 mon.a (mon.0) 500 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:38 vm03 bash[21274]: audit 2026-03-10T10:17:37.257930+0000 mon.a (mon.0) 500 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:38 vm03 bash[21274]: audit 2026-03-10T10:17:37.263199+0000 mon.a (mon.0) 501 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:38 vm03 bash[21274]: audit 2026-03-10T10:17:37.263199+0000 mon.a (mon.0) 501 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:38 vm03 bash[21274]: audit 2026-03-10T10:17:37.268302+0000 mon.a (mon.0) 502 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:38 vm03 bash[21274]: audit 2026-03-10T10:17:37.268302+0000 mon.a (mon.0) 502 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:38 vm03 bash[21274]: audit 2026-03-10T10:17:37.786610+0000 mon.a (mon.0) 503 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:38 vm03 bash[21274]: audit 2026-03-10T10:17:37.786610+0000 mon.a (mon.0) 503 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:38 vm03 bash[21274]: audit 2026-03-10T10:17:37.792332+0000 mon.a (mon.0) 504 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:38 vm03 bash[21274]: audit 2026-03-10T10:17:37.792332+0000 mon.a (mon.0) 504 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:38 vm03 bash[21274]: audit 2026-03-10T10:17:37.797570+0000 mon.a (mon.0) 505 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:38 vm03 bash[21274]: audit 2026-03-10T10:17:37.797570+0000 mon.a (mon.0) 505 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:38 vm00 bash[20709]: audit 2026-03-10T10:17:37.257930+0000 mon.a (mon.0) 500 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:38 vm00 bash[20709]: audit 2026-03-10T10:17:37.257930+0000 mon.a (mon.0) 500 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:38 vm00 bash[20709]: audit 2026-03-10T10:17:37.263199+0000 mon.a (mon.0) 501 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:38 vm00 bash[20709]: audit 2026-03-10T10:17:37.263199+0000 mon.a (mon.0) 501 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:38 vm00 bash[20709]: audit 2026-03-10T10:17:37.268302+0000 mon.a (mon.0) 502 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:38 vm00 bash[20709]: audit 2026-03-10T10:17:37.268302+0000 mon.a (mon.0) 502 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:38 vm00 bash[20709]: audit 2026-03-10T10:17:37.786610+0000 mon.a (mon.0) 503 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:38 vm00 bash[20709]: audit 2026-03-10T10:17:37.786610+0000 mon.a (mon.0) 503 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:38 vm00 bash[20709]: audit 2026-03-10T10:17:37.792332+0000 mon.a (mon.0) 504 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:38 vm00 bash[20709]: audit 2026-03-10T10:17:37.792332+0000 mon.a (mon.0) 504 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:38 vm00 bash[20709]: audit 2026-03-10T10:17:37.797570+0000 mon.a (mon.0) 505 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:38 vm00 bash[20709]: audit 2026-03-10T10:17:37.797570+0000 mon.a (mon.0) 505 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.757 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:38 vm09 bash[20916]: audit 2026-03-10T10:17:37.257930+0000 mon.a (mon.0) 500 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.757 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:38 vm09 bash[20916]: audit 2026-03-10T10:17:37.257930+0000 mon.a (mon.0) 500 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.757 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:38 vm09 bash[20916]: audit 2026-03-10T10:17:37.263199+0000 mon.a (mon.0) 501 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.757 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:38 vm09 bash[20916]: audit 2026-03-10T10:17:37.263199+0000 mon.a (mon.0) 501 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:38 vm09 bash[20916]: audit 2026-03-10T10:17:37.268302+0000 mon.a (mon.0) 502 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:38 vm09 bash[20916]: audit 2026-03-10T10:17:37.268302+0000 mon.a (mon.0) 502 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:38 vm09 bash[20916]: audit 2026-03-10T10:17:37.786610+0000 mon.a (mon.0) 503 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:38 vm09 bash[20916]: audit 2026-03-10T10:17:37.786610+0000 mon.a (mon.0) 503 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:38 vm09 bash[20916]: audit 2026-03-10T10:17:37.792332+0000 mon.a (mon.0) 504 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:38 vm09 bash[20916]: audit 2026-03-10T10:17:37.792332+0000 mon.a (mon.0) 504 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:38 vm09 bash[20916]: audit 2026-03-10T10:17:37.797570+0000 mon.a (mon.0) 505 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:38.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:38 vm09 bash[20916]: audit 2026-03-10T10:17:37.797570+0000 mon.a (mon.0) 505 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:39.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:39 vm09 bash[20916]: cluster 2026-03-10T10:17:37.777681+0000 mgr.a (mgr.14150) 142 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:39.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:39 vm09 bash[20916]: cluster 2026-03-10T10:17:37.777681+0000 mgr.a (mgr.14150) 142 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:39.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:39 vm09 bash[20916]: audit 2026-03-10T10:17:39.266298+0000 mon.a (mon.0) 506 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:39.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:39 vm09 bash[20916]: audit 2026-03-10T10:17:39.266298+0000 mon.a (mon.0) 506 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:39.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:39 vm09 bash[20916]: audit 2026-03-10T10:17:39.269972+0000 mon.a (mon.0) 507 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:39.507 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:39 vm09 bash[20916]: audit 2026-03-10T10:17:39.269972+0000 mon.a (mon.0) 507 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:39.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:39 vm03 bash[21274]: cluster 2026-03-10T10:17:37.777681+0000 mgr.a (mgr.14150) 142 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:39.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:39 vm03 bash[21274]: cluster 2026-03-10T10:17:37.777681+0000 mgr.a (mgr.14150) 142 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:39.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:39 vm03 bash[21274]: audit 2026-03-10T10:17:39.266298+0000 mon.a (mon.0) 506 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:39.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:39 vm03 bash[21274]: audit 2026-03-10T10:17:39.266298+0000 mon.a (mon.0) 506 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:39.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:39 vm03 bash[21274]: audit 2026-03-10T10:17:39.269972+0000 mon.a (mon.0) 507 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:39.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:39 vm03 bash[21274]: audit 2026-03-10T10:17:39.269972+0000 mon.a (mon.0) 507 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:39.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:39 vm00 bash[20709]: cluster 2026-03-10T10:17:37.777681+0000 mgr.a (mgr.14150) 142 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:39.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:39 vm00 bash[20709]: cluster 2026-03-10T10:17:37.777681+0000 mgr.a (mgr.14150) 142 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:39.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:39 vm00 bash[20709]: audit 2026-03-10T10:17:39.266298+0000 mon.a (mon.0) 506 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:39.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:39 vm00 bash[20709]: audit 2026-03-10T10:17:39.266298+0000 mon.a (mon.0) 506 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:39.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:39 vm00 bash[20709]: audit 2026-03-10T10:17:39.269972+0000 mon.a (mon.0) 507 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:39.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:39 vm00 bash[20709]: audit 2026-03-10T10:17:39.269972+0000 mon.a (mon.0) 507 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:40.465 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:40 vm09 bash[20916]: audit 2026-03-10T10:17:39.273869+0000 mon.a (mon.0) 508 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:40.465 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:40 vm09 bash[20916]: audit 2026-03-10T10:17:39.273869+0000 mon.a (mon.0) 508 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:40.465 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:40 vm09 bash[20916]: audit 2026-03-10T10:17:39.647171+0000 mon.a (mon.0) 509 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T10:17:40.465 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:40 vm09 bash[20916]: audit 2026-03-10T10:17:39.647171+0000 mon.a (mon.0) 509 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T10:17:40.465 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:40 vm09 bash[20916]: audit 2026-03-10T10:17:39.647688+0000 mon.a (mon.0) 510 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:40.465 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:40 vm09 bash[20916]: audit 2026-03-10T10:17:39.647688+0000 mon.a (mon.0) 510 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:40.465 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:40 vm09 bash[20916]: cephadm 2026-03-10T10:17:39.648096+0000 mgr.a (mgr.14150) 143 : cephadm [INF] Deploying daemon osd.2 on vm09 2026-03-10T10:17:40.465 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:40 vm09 bash[20916]: cephadm 2026-03-10T10:17:39.648096+0000 mgr.a (mgr.14150) 143 : cephadm [INF] Deploying daemon osd.2 on vm09 2026-03-10T10:17:40.465 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:40 vm09 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:17:40.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:40 vm03 bash[21274]: audit 2026-03-10T10:17:39.273869+0000 mon.a (mon.0) 508 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:40.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:40 vm03 bash[21274]: audit 2026-03-10T10:17:39.273869+0000 mon.a (mon.0) 508 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:40.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:40 vm03 bash[21274]: audit 2026-03-10T10:17:39.647171+0000 mon.a (mon.0) 509 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T10:17:40.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:40 vm03 bash[21274]: audit 2026-03-10T10:17:39.647171+0000 mon.a (mon.0) 509 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T10:17:40.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:40 vm03 bash[21274]: audit 2026-03-10T10:17:39.647688+0000 mon.a (mon.0) 510 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:40.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:40 vm03 bash[21274]: audit 2026-03-10T10:17:39.647688+0000 mon.a (mon.0) 510 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:40.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:40 vm03 bash[21274]: cephadm 2026-03-10T10:17:39.648096+0000 mgr.a (mgr.14150) 143 : cephadm [INF] Deploying daemon osd.2 on vm09 2026-03-10T10:17:40.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:40 vm03 bash[21274]: cephadm 2026-03-10T10:17:39.648096+0000 mgr.a (mgr.14150) 143 : cephadm [INF] Deploying daemon osd.2 on vm09 2026-03-10T10:17:40.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:40 vm00 bash[20709]: audit 2026-03-10T10:17:39.273869+0000 mon.a (mon.0) 508 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:40.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:40 vm00 bash[20709]: audit 2026-03-10T10:17:39.273869+0000 mon.a (mon.0) 508 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:40.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:40 vm00 bash[20709]: audit 2026-03-10T10:17:39.647171+0000 mon.a (mon.0) 509 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T10:17:40.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:40 vm00 bash[20709]: audit 2026-03-10T10:17:39.647171+0000 mon.a (mon.0) 509 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T10:17:40.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:40 vm00 bash[20709]: audit 2026-03-10T10:17:39.647688+0000 mon.a (mon.0) 510 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:40.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:40 vm00 bash[20709]: audit 2026-03-10T10:17:39.647688+0000 mon.a (mon.0) 510 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:40.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:40 vm00 bash[20709]: cephadm 2026-03-10T10:17:39.648096+0000 mgr.a (mgr.14150) 143 : cephadm [INF] Deploying daemon osd.2 on vm09 2026-03-10T10:17:40.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:40 vm00 bash[20709]: cephadm 2026-03-10T10:17:39.648096+0000 mgr.a (mgr.14150) 143 : cephadm [INF] Deploying daemon osd.2 on vm09 2026-03-10T10:17:40.753 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:40 vm09 systemd[1]: /etc/systemd/system/ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-10T10:17:41.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:41 vm03 bash[21274]: cluster 2026-03-10T10:17:39.777906+0000 mgr.a (mgr.14150) 144 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:41.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:41 vm03 bash[21274]: cluster 2026-03-10T10:17:39.777906+0000 mgr.a (mgr.14150) 144 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:41.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:41 vm03 bash[21274]: audit 2026-03-10T10:17:40.776479+0000 mon.a (mon.0) 511 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:41.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:41 vm03 bash[21274]: audit 2026-03-10T10:17:40.776479+0000 mon.a (mon.0) 511 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:41.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:41 vm03 bash[21274]: audit 2026-03-10T10:17:40.777883+0000 mon.a (mon.0) 512 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:41.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:41 vm03 bash[21274]: audit 2026-03-10T10:17:40.777883+0000 mon.a (mon.0) 512 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:41.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:41 vm03 bash[21274]: audit 2026-03-10T10:17:40.778342+0000 mon.a (mon.0) 513 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:41.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:41 vm03 bash[21274]: audit 2026-03-10T10:17:40.778342+0000 mon.a (mon.0) 513 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:41.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:41 vm03 bash[21274]: audit 2026-03-10T10:17:40.781459+0000 mon.a (mon.0) 514 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:41.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:41 vm03 bash[21274]: audit 2026-03-10T10:17:40.781459+0000 mon.a (mon.0) 514 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:41.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:41 vm03 bash[21274]: audit 2026-03-10T10:17:40.784858+0000 mon.a (mon.0) 515 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:41.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:41 vm03 bash[21274]: audit 2026-03-10T10:17:40.784858+0000 mon.a (mon.0) 515 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:41.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:41 vm03 bash[21274]: audit 2026-03-10T10:17:40.787935+0000 mon.a (mon.0) 516 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:41.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:41 vm03 bash[21274]: audit 2026-03-10T10:17:40.787935+0000 mon.a (mon.0) 516 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:41.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:41 vm00 bash[20709]: cluster 2026-03-10T10:17:39.777906+0000 mgr.a (mgr.14150) 144 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:41.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:41 vm00 bash[20709]: cluster 2026-03-10T10:17:39.777906+0000 mgr.a (mgr.14150) 144 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:41.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:41 vm00 bash[20709]: audit 2026-03-10T10:17:40.776479+0000 mon.a (mon.0) 511 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:41.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:41 vm00 bash[20709]: audit 2026-03-10T10:17:40.776479+0000 mon.a (mon.0) 511 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:41.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:41 vm00 bash[20709]: audit 2026-03-10T10:17:40.777883+0000 mon.a (mon.0) 512 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:41.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:41 vm00 bash[20709]: audit 2026-03-10T10:17:40.777883+0000 mon.a (mon.0) 512 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:41.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:41 vm00 bash[20709]: audit 2026-03-10T10:17:40.778342+0000 mon.a (mon.0) 513 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:41.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:41 vm00 bash[20709]: audit 2026-03-10T10:17:40.778342+0000 mon.a (mon.0) 513 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:41.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:41 vm00 bash[20709]: audit 2026-03-10T10:17:40.781459+0000 mon.a (mon.0) 514 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:41.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:41 vm00 bash[20709]: audit 2026-03-10T10:17:40.781459+0000 mon.a (mon.0) 514 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:41.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:41 vm00 bash[20709]: audit 2026-03-10T10:17:40.784858+0000 mon.a (mon.0) 515 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:41.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:41 vm00 bash[20709]: audit 2026-03-10T10:17:40.784858+0000 mon.a (mon.0) 515 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:41.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:41 vm00 bash[20709]: audit 2026-03-10T10:17:40.787935+0000 mon.a (mon.0) 516 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:41.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:41 vm00 bash[20709]: audit 2026-03-10T10:17:40.787935+0000 mon.a (mon.0) 516 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:41.757 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:41 vm09 bash[20916]: cluster 2026-03-10T10:17:39.777906+0000 mgr.a (mgr.14150) 144 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:41.757 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:41 vm09 bash[20916]: cluster 2026-03-10T10:17:39.777906+0000 mgr.a (mgr.14150) 144 : cluster [DBG] pgmap v82: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:41.757 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:41 vm09 bash[20916]: audit 2026-03-10T10:17:40.776479+0000 mon.a (mon.0) 511 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:41.757 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:41 vm09 bash[20916]: audit 2026-03-10T10:17:40.776479+0000 mon.a (mon.0) 511 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:41.757 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:41 vm09 bash[20916]: audit 2026-03-10T10:17:40.777883+0000 mon.a (mon.0) 512 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:41.757 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:41 vm09 bash[20916]: audit 2026-03-10T10:17:40.777883+0000 mon.a (mon.0) 512 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:41.757 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:41 vm09 bash[20916]: audit 2026-03-10T10:17:40.778342+0000 mon.a (mon.0) 513 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:41.757 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:41 vm09 bash[20916]: audit 2026-03-10T10:17:40.778342+0000 mon.a (mon.0) 513 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:41.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:41 vm09 bash[20916]: audit 2026-03-10T10:17:40.781459+0000 mon.a (mon.0) 514 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:41.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:41 vm09 bash[20916]: audit 2026-03-10T10:17:40.781459+0000 mon.a (mon.0) 514 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:41.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:41 vm09 bash[20916]: audit 2026-03-10T10:17:40.784858+0000 mon.a (mon.0) 515 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:41.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:41 vm09 bash[20916]: audit 2026-03-10T10:17:40.784858+0000 mon.a (mon.0) 515 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:41.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:41 vm09 bash[20916]: audit 2026-03-10T10:17:40.787935+0000 mon.a (mon.0) 516 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:41.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:41 vm09 bash[20916]: audit 2026-03-10T10:17:40.787935+0000 mon.a (mon.0) 516 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:43.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:43 vm00 bash[20709]: cluster 2026-03-10T10:17:41.778090+0000 mgr.a (mgr.14150) 145 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:43.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:43 vm00 bash[20709]: cluster 2026-03-10T10:17:41.778090+0000 mgr.a (mgr.14150) 145 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:43.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:43 vm03 bash[21274]: cluster 2026-03-10T10:17:41.778090+0000 mgr.a (mgr.14150) 145 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:43.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:43 vm03 bash[21274]: cluster 2026-03-10T10:17:41.778090+0000 mgr.a (mgr.14150) 145 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:44.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:43 vm09 bash[20916]: cluster 2026-03-10T10:17:41.778090+0000 mgr.a (mgr.14150) 145 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:44.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:43 vm09 bash[20916]: cluster 2026-03-10T10:17:41.778090+0000 mgr.a (mgr.14150) 145 : cluster [DBG] pgmap v83: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:44.411 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:44 vm09 bash[20916]: audit 2026-03-10T10:17:44.019424+0000 mon.c (mon.1) 10 : audit [INF] from='osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T10:17:44.411 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:44 vm09 bash[20916]: audit 2026-03-10T10:17:44.019424+0000 mon.c (mon.1) 10 : audit [INF] from='osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T10:17:44.411 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:44 vm09 bash[20916]: audit 2026-03-10T10:17:44.020019+0000 mon.a (mon.0) 517 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T10:17:44.411 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:44 vm09 bash[20916]: audit 2026-03-10T10:17:44.020019+0000 mon.a (mon.0) 517 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T10:17:44.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:44 vm00 bash[20709]: audit 2026-03-10T10:17:44.019424+0000 mon.c (mon.1) 10 : audit [INF] from='osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T10:17:44.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:44 vm00 bash[20709]: audit 2026-03-10T10:17:44.019424+0000 mon.c (mon.1) 10 : audit [INF] from='osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T10:17:44.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:44 vm00 bash[20709]: audit 2026-03-10T10:17:44.020019+0000 mon.a (mon.0) 517 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T10:17:44.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:44 vm00 bash[20709]: audit 2026-03-10T10:17:44.020019+0000 mon.a (mon.0) 517 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T10:17:44.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:44 vm03 bash[21274]: audit 2026-03-10T10:17:44.019424+0000 mon.c (mon.1) 10 : audit [INF] from='osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T10:17:44.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:44 vm03 bash[21274]: audit 2026-03-10T10:17:44.019424+0000 mon.c (mon.1) 10 : audit [INF] from='osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T10:17:44.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:44 vm03 bash[21274]: audit 2026-03-10T10:17:44.020019+0000 mon.a (mon.0) 517 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T10:17:44.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:44 vm03 bash[21274]: audit 2026-03-10T10:17:44.020019+0000 mon.a (mon.0) 517 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T10:17:45.753 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:45 vm09 bash[20916]: cluster 2026-03-10T10:17:43.778380+0000 mgr.a (mgr.14150) 146 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:45.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:45 vm09 bash[20916]: cluster 2026-03-10T10:17:43.778380+0000 mgr.a (mgr.14150) 146 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:45.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:45 vm09 bash[20916]: audit 2026-03-10T10:17:44.417172+0000 mon.a (mon.0) 518 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T10:17:45.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:45 vm09 bash[20916]: audit 2026-03-10T10:17:44.417172+0000 mon.a (mon.0) 518 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T10:17:45.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:45 vm09 bash[20916]: cluster 2026-03-10T10:17:44.419426+0000 mon.a (mon.0) 519 : cluster [DBG] osdmap e16: 3 total, 2 up, 3 in 2026-03-10T10:17:45.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:45 vm09 bash[20916]: cluster 2026-03-10T10:17:44.419426+0000 mon.a (mon.0) 519 : cluster [DBG] osdmap e16: 3 total, 2 up, 3 in 2026-03-10T10:17:45.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:45 vm09 bash[20916]: audit 2026-03-10T10:17:44.419670+0000 mon.a (mon.0) 520 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:45.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:45 vm09 bash[20916]: audit 2026-03-10T10:17:44.419670+0000 mon.a (mon.0) 520 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:45.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:45 vm09 bash[20916]: audit 2026-03-10T10:17:44.420809+0000 mon.c (mon.1) 11 : audit [INF] from='osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:17:45.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:45 vm09 bash[20916]: audit 2026-03-10T10:17:44.420809+0000 mon.c (mon.1) 11 : audit [INF] from='osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:17:45.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:45 vm09 bash[20916]: audit 2026-03-10T10:17:44.421305+0000 mon.a (mon.0) 521 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:17:45.754 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:45 vm09 bash[20916]: audit 2026-03-10T10:17:44.421305+0000 mon.a (mon.0) 521 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:17:45.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:45 vm03 bash[21274]: cluster 2026-03-10T10:17:43.778380+0000 mgr.a (mgr.14150) 146 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:45 vm03 bash[21274]: cluster 2026-03-10T10:17:43.778380+0000 mgr.a (mgr.14150) 146 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:45 vm03 bash[21274]: audit 2026-03-10T10:17:44.417172+0000 mon.a (mon.0) 518 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T10:17:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:45 vm03 bash[21274]: audit 2026-03-10T10:17:44.417172+0000 mon.a (mon.0) 518 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T10:17:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:45 vm03 bash[21274]: cluster 2026-03-10T10:17:44.419426+0000 mon.a (mon.0) 519 : cluster [DBG] osdmap e16: 3 total, 2 up, 3 in 2026-03-10T10:17:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:45 vm03 bash[21274]: cluster 2026-03-10T10:17:44.419426+0000 mon.a (mon.0) 519 : cluster [DBG] osdmap e16: 3 total, 2 up, 3 in 2026-03-10T10:17:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:45 vm03 bash[21274]: audit 2026-03-10T10:17:44.419670+0000 mon.a (mon.0) 520 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:45 vm03 bash[21274]: audit 2026-03-10T10:17:44.419670+0000 mon.a (mon.0) 520 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:45 vm03 bash[21274]: audit 2026-03-10T10:17:44.420809+0000 mon.c (mon.1) 11 : audit [INF] from='osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:17:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:45 vm03 bash[21274]: audit 2026-03-10T10:17:44.420809+0000 mon.c (mon.1) 11 : audit [INF] from='osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:17:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:45 vm03 bash[21274]: audit 2026-03-10T10:17:44.421305+0000 mon.a (mon.0) 521 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:17:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:45 vm03 bash[21274]: audit 2026-03-10T10:17:44.421305+0000 mon.a (mon.0) 521 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:17:45.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:45 vm00 bash[20709]: cluster 2026-03-10T10:17:43.778380+0000 mgr.a (mgr.14150) 146 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:45.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:45 vm00 bash[20709]: cluster 2026-03-10T10:17:43.778380+0000 mgr.a (mgr.14150) 146 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:45.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:45 vm00 bash[20709]: audit 2026-03-10T10:17:44.417172+0000 mon.a (mon.0) 518 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T10:17:45.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:45 vm00 bash[20709]: audit 2026-03-10T10:17:44.417172+0000 mon.a (mon.0) 518 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T10:17:45.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:45 vm00 bash[20709]: cluster 2026-03-10T10:17:44.419426+0000 mon.a (mon.0) 519 : cluster [DBG] osdmap e16: 3 total, 2 up, 3 in 2026-03-10T10:17:45.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:45 vm00 bash[20709]: cluster 2026-03-10T10:17:44.419426+0000 mon.a (mon.0) 519 : cluster [DBG] osdmap e16: 3 total, 2 up, 3 in 2026-03-10T10:17:45.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:45 vm00 bash[20709]: audit 2026-03-10T10:17:44.419670+0000 mon.a (mon.0) 520 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:45.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:45 vm00 bash[20709]: audit 2026-03-10T10:17:44.419670+0000 mon.a (mon.0) 520 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:45.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:45 vm00 bash[20709]: audit 2026-03-10T10:17:44.420809+0000 mon.c (mon.1) 11 : audit [INF] from='osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:17:45.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:45 vm00 bash[20709]: audit 2026-03-10T10:17:44.420809+0000 mon.c (mon.1) 11 : audit [INF] from='osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:17:45.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:45 vm00 bash[20709]: audit 2026-03-10T10:17:44.421305+0000 mon.a (mon.0) 521 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:17:45.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:45 vm00 bash[20709]: audit 2026-03-10T10:17:44.421305+0000 mon.a (mon.0) 521 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:17:46.707 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:46 vm09 bash[20916]: audit 2026-03-10T10:17:45.420594+0000 mon.a (mon.0) 522 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T10:17:46.708 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:46 vm09 bash[20916]: audit 2026-03-10T10:17:45.420594+0000 mon.a (mon.0) 522 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T10:17:46.708 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:46 vm09 bash[20916]: cluster 2026-03-10T10:17:45.423463+0000 mon.a (mon.0) 523 : cluster [DBG] osdmap e17: 3 total, 2 up, 3 in 2026-03-10T10:17:46.708 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:46 vm09 bash[20916]: cluster 2026-03-10T10:17:45.423463+0000 mon.a (mon.0) 523 : cluster [DBG] osdmap e17: 3 total, 2 up, 3 in 2026-03-10T10:17:46.708 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:46 vm09 bash[20916]: audit 2026-03-10T10:17:45.424137+0000 mon.a (mon.0) 524 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:46.708 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:46 vm09 bash[20916]: audit 2026-03-10T10:17:45.424137+0000 mon.a (mon.0) 524 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:46.708 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:46 vm09 bash[20916]: audit 2026-03-10T10:17:45.431442+0000 mon.a (mon.0) 525 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:46.708 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:46 vm09 bash[20916]: audit 2026-03-10T10:17:45.431442+0000 mon.a (mon.0) 525 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:46.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:46 vm03 bash[21274]: audit 2026-03-10T10:17:45.420594+0000 mon.a (mon.0) 522 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T10:17:46.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:46 vm03 bash[21274]: audit 2026-03-10T10:17:45.420594+0000 mon.a (mon.0) 522 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T10:17:46.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:46 vm03 bash[21274]: cluster 2026-03-10T10:17:45.423463+0000 mon.a (mon.0) 523 : cluster [DBG] osdmap e17: 3 total, 2 up, 3 in 2026-03-10T10:17:46.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:46 vm03 bash[21274]: cluster 2026-03-10T10:17:45.423463+0000 mon.a (mon.0) 523 : cluster [DBG] osdmap e17: 3 total, 2 up, 3 in 2026-03-10T10:17:46.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:46 vm03 bash[21274]: audit 2026-03-10T10:17:45.424137+0000 mon.a (mon.0) 524 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:46.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:46 vm03 bash[21274]: audit 2026-03-10T10:17:45.424137+0000 mon.a (mon.0) 524 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:46.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:46 vm03 bash[21274]: audit 2026-03-10T10:17:45.431442+0000 mon.a (mon.0) 525 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:46.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:46 vm03 bash[21274]: audit 2026-03-10T10:17:45.431442+0000 mon.a (mon.0) 525 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:46.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:46 vm00 bash[20709]: audit 2026-03-10T10:17:45.420594+0000 mon.a (mon.0) 522 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T10:17:46.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:46 vm00 bash[20709]: audit 2026-03-10T10:17:45.420594+0000 mon.a (mon.0) 522 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T10:17:46.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:46 vm00 bash[20709]: cluster 2026-03-10T10:17:45.423463+0000 mon.a (mon.0) 523 : cluster [DBG] osdmap e17: 3 total, 2 up, 3 in 2026-03-10T10:17:46.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:46 vm00 bash[20709]: cluster 2026-03-10T10:17:45.423463+0000 mon.a (mon.0) 523 : cluster [DBG] osdmap e17: 3 total, 2 up, 3 in 2026-03-10T10:17:46.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:46 vm00 bash[20709]: audit 2026-03-10T10:17:45.424137+0000 mon.a (mon.0) 524 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:46.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:46 vm00 bash[20709]: audit 2026-03-10T10:17:45.424137+0000 mon.a (mon.0) 524 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:46.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:46 vm00 bash[20709]: audit 2026-03-10T10:17:45.431442+0000 mon.a (mon.0) 525 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:46.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:46 vm00 bash[20709]: audit 2026-03-10T10:17:45.431442+0000 mon.a (mon.0) 525 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:47.712 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:47 vm09 bash[20916]: cluster 2026-03-10T10:17:45.001735+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:17:47.713 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:47 vm09 bash[20916]: cluster 2026-03-10T10:17:45.001735+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:17:47.713 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:47 vm09 bash[20916]: cluster 2026-03-10T10:17:45.001791+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:17:47.713 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:47 vm09 bash[20916]: cluster 2026-03-10T10:17:45.001791+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:17:47.713 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:47 vm09 bash[20916]: cluster 2026-03-10T10:17:45.778612+0000 mgr.a (mgr.14150) 147 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:47.713 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:47 vm09 bash[20916]: cluster 2026-03-10T10:17:45.778612+0000 mgr.a (mgr.14150) 147 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:47.713 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:47 vm09 bash[20916]: audit 2026-03-10T10:17:46.429668+0000 mon.a (mon.0) 526 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:47.713 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:47 vm09 bash[20916]: audit 2026-03-10T10:17:46.429668+0000 mon.a (mon.0) 526 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:47.713 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:47 vm09 bash[20916]: cluster 2026-03-10T10:17:46.434872+0000 mon.a (mon.0) 527 : cluster [INF] osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425] boot 2026-03-10T10:17:47.713 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:47 vm09 bash[20916]: cluster 2026-03-10T10:17:46.434872+0000 mon.a (mon.0) 527 : cluster [INF] osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425] boot 2026-03-10T10:17:47.713 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:47 vm09 bash[20916]: cluster 2026-03-10T10:17:46.434938+0000 mon.a (mon.0) 528 : cluster [DBG] osdmap e18: 3 total, 3 up, 3 in 2026-03-10T10:17:47.713 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:47 vm09 bash[20916]: cluster 2026-03-10T10:17:46.434938+0000 mon.a (mon.0) 528 : cluster [DBG] osdmap e18: 3 total, 3 up, 3 in 2026-03-10T10:17:47.713 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:47 vm09 bash[20916]: audit 2026-03-10T10:17:46.435794+0000 mon.a (mon.0) 529 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:47.713 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:47 vm09 bash[20916]: audit 2026-03-10T10:17:46.435794+0000 mon.a (mon.0) 529 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:47.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:47 vm03 bash[21274]: cluster 2026-03-10T10:17:45.001735+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:17:47.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:47 vm03 bash[21274]: cluster 2026-03-10T10:17:45.001735+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:17:47.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:47 vm03 bash[21274]: cluster 2026-03-10T10:17:45.001791+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:17:47.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:47 vm03 bash[21274]: cluster 2026-03-10T10:17:45.001791+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:17:47.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:47 vm03 bash[21274]: cluster 2026-03-10T10:17:45.778612+0000 mgr.a (mgr.14150) 147 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:47.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:47 vm03 bash[21274]: cluster 2026-03-10T10:17:45.778612+0000 mgr.a (mgr.14150) 147 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:47.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:47 vm03 bash[21274]: audit 2026-03-10T10:17:46.429668+0000 mon.a (mon.0) 526 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:47.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:47 vm03 bash[21274]: audit 2026-03-10T10:17:46.429668+0000 mon.a (mon.0) 526 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:47.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:47 vm03 bash[21274]: cluster 2026-03-10T10:17:46.434872+0000 mon.a (mon.0) 527 : cluster [INF] osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425] boot 2026-03-10T10:17:47.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:47 vm03 bash[21274]: cluster 2026-03-10T10:17:46.434872+0000 mon.a (mon.0) 527 : cluster [INF] osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425] boot 2026-03-10T10:17:47.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:47 vm03 bash[21274]: cluster 2026-03-10T10:17:46.434938+0000 mon.a (mon.0) 528 : cluster [DBG] osdmap e18: 3 total, 3 up, 3 in 2026-03-10T10:17:47.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:47 vm03 bash[21274]: cluster 2026-03-10T10:17:46.434938+0000 mon.a (mon.0) 528 : cluster [DBG] osdmap e18: 3 total, 3 up, 3 in 2026-03-10T10:17:47.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:47 vm03 bash[21274]: audit 2026-03-10T10:17:46.435794+0000 mon.a (mon.0) 529 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:47.835 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:47 vm03 bash[21274]: audit 2026-03-10T10:17:46.435794+0000 mon.a (mon.0) 529 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:47.890 INFO:teuthology.orchestra.run.vm09.stdout:Created osd(s) 2 on host 'vm09' 2026-03-10T10:17:47.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:47 vm00 bash[20709]: cluster 2026-03-10T10:17:45.001735+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:17:47.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:47 vm00 bash[20709]: cluster 2026-03-10T10:17:45.001735+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-10T10:17:47.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:47 vm00 bash[20709]: cluster 2026-03-10T10:17:45.001791+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:17:47.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:47 vm00 bash[20709]: cluster 2026-03-10T10:17:45.001791+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-10T10:17:47.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:47 vm00 bash[20709]: cluster 2026-03-10T10:17:45.778612+0000 mgr.a (mgr.14150) 147 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:47.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:47 vm00 bash[20709]: cluster 2026-03-10T10:17:45.778612+0000 mgr.a (mgr.14150) 147 : cluster [DBG] pgmap v87: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:17:47.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:47 vm00 bash[20709]: audit 2026-03-10T10:17:46.429668+0000 mon.a (mon.0) 526 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:47.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:47 vm00 bash[20709]: audit 2026-03-10T10:17:46.429668+0000 mon.a (mon.0) 526 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:47.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:47 vm00 bash[20709]: cluster 2026-03-10T10:17:46.434872+0000 mon.a (mon.0) 527 : cluster [INF] osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425] boot 2026-03-10T10:17:47.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:47 vm00 bash[20709]: cluster 2026-03-10T10:17:46.434872+0000 mon.a (mon.0) 527 : cluster [INF] osd.2 [v2:192.168.123.109:6800/2811766425,v1:192.168.123.109:6801/2811766425] boot 2026-03-10T10:17:47.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:47 vm00 bash[20709]: cluster 2026-03-10T10:17:46.434938+0000 mon.a (mon.0) 528 : cluster [DBG] osdmap e18: 3 total, 3 up, 3 in 2026-03-10T10:17:47.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:47 vm00 bash[20709]: cluster 2026-03-10T10:17:46.434938+0000 mon.a (mon.0) 528 : cluster [DBG] osdmap e18: 3 total, 3 up, 3 in 2026-03-10T10:17:47.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:47 vm00 bash[20709]: audit 2026-03-10T10:17:46.435794+0000 mon.a (mon.0) 529 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:47.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:47 vm00 bash[20709]: audit 2026-03-10T10:17:46.435794+0000 mon.a (mon.0) 529 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:17:47.950 DEBUG:teuthology.orchestra.run.vm09:osd.2> sudo journalctl -f -n 0 -u ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@osd.2.service 2026-03-10T10:17:47.951 INFO:tasks.cephadm:Waiting for 3 OSDs to come up... 2026-03-10T10:17:47.952 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph osd stat -f json 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: cluster 2026-03-10T10:17:47.455758+0000 mon.a (mon.0) 530 : cluster [DBG] osdmap e19: 3 total, 3 up, 3 in 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: cluster 2026-03-10T10:17:47.455758+0000 mon.a (mon.0) 530 : cluster [DBG] osdmap e19: 3 total, 3 up, 3 in 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.761964+0000 mon.a (mon.0) 531 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.761964+0000 mon.a (mon.0) 531 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.764347+0000 mon.a (mon.0) 532 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.764347+0000 mon.a (mon.0) 532 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.768033+0000 mon.a (mon.0) 533 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.768033+0000 mon.a (mon.0) 533 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.842020+0000 mon.a (mon.0) 534 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.842020+0000 mon.a (mon.0) 534 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.886105+0000 mon.a (mon.0) 535 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.886105+0000 mon.a (mon.0) 535 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.890370+0000 mon.a (mon.0) 536 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.890370+0000 mon.a (mon.0) 536 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.892079+0000 mon.a (mon.0) 537 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.892079+0000 mon.a (mon.0) 537 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.892632+0000 mon.a (mon.0) 538 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.892632+0000 mon.a (mon.0) 538 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.896770+0000 mon.a (mon.0) 539 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:48 vm09 bash[20916]: audit 2026-03-10T10:17:47.896770+0000 mon.a (mon.0) 539 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: cluster 2026-03-10T10:17:47.455758+0000 mon.a (mon.0) 530 : cluster [DBG] osdmap e19: 3 total, 3 up, 3 in 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: cluster 2026-03-10T10:17:47.455758+0000 mon.a (mon.0) 530 : cluster [DBG] osdmap e19: 3 total, 3 up, 3 in 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.761964+0000 mon.a (mon.0) 531 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.761964+0000 mon.a (mon.0) 531 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.764347+0000 mon.a (mon.0) 532 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.764347+0000 mon.a (mon.0) 532 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.768033+0000 mon.a (mon.0) 533 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.768033+0000 mon.a (mon.0) 533 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.842020+0000 mon.a (mon.0) 534 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.842020+0000 mon.a (mon.0) 534 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.886105+0000 mon.a (mon.0) 535 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.886105+0000 mon.a (mon.0) 535 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.890370+0000 mon.a (mon.0) 536 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.890370+0000 mon.a (mon.0) 536 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.892079+0000 mon.a (mon.0) 537 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.892079+0000 mon.a (mon.0) 537 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.892632+0000 mon.a (mon.0) 538 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.892632+0000 mon.a (mon.0) 538 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.896770+0000 mon.a (mon.0) 539 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:48 vm03 bash[21274]: audit 2026-03-10T10:17:47.896770+0000 mon.a (mon.0) 539 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: cluster 2026-03-10T10:17:47.455758+0000 mon.a (mon.0) 530 : cluster [DBG] osdmap e19: 3 total, 3 up, 3 in 2026-03-10T10:17:48.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: cluster 2026-03-10T10:17:47.455758+0000 mon.a (mon.0) 530 : cluster [DBG] osdmap e19: 3 total, 3 up, 3 in 2026-03-10T10:17:48.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.761964+0000 mon.a (mon.0) 531 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:48.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.761964+0000 mon.a (mon.0) 531 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:48.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.764347+0000 mon.a (mon.0) 532 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:17:48.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.764347+0000 mon.a (mon.0) 532 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:17:48.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.768033+0000 mon.a (mon.0) 533 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.768033+0000 mon.a (mon.0) 533 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.842020+0000 mon.a (mon.0) 534 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:17:48.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.842020+0000 mon.a (mon.0) 534 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:17:48.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.886105+0000 mon.a (mon.0) 535 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.886105+0000 mon.a (mon.0) 535 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.890370+0000 mon.a (mon.0) 536 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.890370+0000 mon.a (mon.0) 536 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.892079+0000 mon.a (mon.0) 537 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:48.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.892079+0000 mon.a (mon.0) 537 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:48.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.892632+0000 mon.a (mon.0) 538 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:48.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.892632+0000 mon.a (mon.0) 538 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:48.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.896770+0000 mon.a (mon.0) 539 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:48.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:48 vm00 bash[20709]: audit 2026-03-10T10:17:47.896770+0000 mon.a (mon.0) 539 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:49.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:49 vm09 bash[20916]: cephadm 2026-03-10T10:17:47.768853+0000 mgr.a (mgr.14150) 148 : cephadm [INF] Adjusting osd_memory_target on vm09 to 3527M 2026-03-10T10:17:49.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:49 vm09 bash[20916]: cephadm 2026-03-10T10:17:47.768853+0000 mgr.a (mgr.14150) 148 : cephadm [INF] Adjusting osd_memory_target on vm09 to 3527M 2026-03-10T10:17:49.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:49 vm09 bash[20916]: cluster 2026-03-10T10:17:47.778802+0000 mgr.a (mgr.14150) 149 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:49.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:49 vm09 bash[20916]: cluster 2026-03-10T10:17:47.778802+0000 mgr.a (mgr.14150) 149 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:49.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:49 vm09 bash[20916]: audit 2026-03-10T10:17:48.773335+0000 mon.a (mon.0) 540 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T10:17:49.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:49 vm09 bash[20916]: audit 2026-03-10T10:17:48.773335+0000 mon.a (mon.0) 540 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T10:17:49.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:49 vm09 bash[20916]: cluster 2026-03-10T10:17:48.775840+0000 mon.a (mon.0) 541 : cluster [DBG] osdmap e20: 3 total, 3 up, 3 in 2026-03-10T10:17:49.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:49 vm09 bash[20916]: cluster 2026-03-10T10:17:48.775840+0000 mon.a (mon.0) 541 : cluster [DBG] osdmap e20: 3 total, 3 up, 3 in 2026-03-10T10:17:49.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:49 vm09 bash[20916]: audit 2026-03-10T10:17:48.776664+0000 mon.a (mon.0) 542 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:17:49.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:49 vm09 bash[20916]: audit 2026-03-10T10:17:48.776664+0000 mon.a (mon.0) 542 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:17:49.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:49 vm03 bash[21274]: cephadm 2026-03-10T10:17:47.768853+0000 mgr.a (mgr.14150) 148 : cephadm [INF] Adjusting osd_memory_target on vm09 to 3527M 2026-03-10T10:17:49.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:49 vm03 bash[21274]: cephadm 2026-03-10T10:17:47.768853+0000 mgr.a (mgr.14150) 148 : cephadm [INF] Adjusting osd_memory_target on vm09 to 3527M 2026-03-10T10:17:49.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:49 vm03 bash[21274]: cluster 2026-03-10T10:17:47.778802+0000 mgr.a (mgr.14150) 149 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:49.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:49 vm03 bash[21274]: cluster 2026-03-10T10:17:47.778802+0000 mgr.a (mgr.14150) 149 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:49.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:49 vm03 bash[21274]: audit 2026-03-10T10:17:48.773335+0000 mon.a (mon.0) 540 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T10:17:49.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:49 vm03 bash[21274]: audit 2026-03-10T10:17:48.773335+0000 mon.a (mon.0) 540 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T10:17:49.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:49 vm03 bash[21274]: cluster 2026-03-10T10:17:48.775840+0000 mon.a (mon.0) 541 : cluster [DBG] osdmap e20: 3 total, 3 up, 3 in 2026-03-10T10:17:49.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:49 vm03 bash[21274]: cluster 2026-03-10T10:17:48.775840+0000 mon.a (mon.0) 541 : cluster [DBG] osdmap e20: 3 total, 3 up, 3 in 2026-03-10T10:17:49.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:49 vm03 bash[21274]: audit 2026-03-10T10:17:48.776664+0000 mon.a (mon.0) 542 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:17:49.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:49 vm03 bash[21274]: audit 2026-03-10T10:17:48.776664+0000 mon.a (mon.0) 542 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:17:49.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:49 vm00 bash[20709]: cephadm 2026-03-10T10:17:47.768853+0000 mgr.a (mgr.14150) 148 : cephadm [INF] Adjusting osd_memory_target on vm09 to 3527M 2026-03-10T10:17:49.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:49 vm00 bash[20709]: cephadm 2026-03-10T10:17:47.768853+0000 mgr.a (mgr.14150) 148 : cephadm [INF] Adjusting osd_memory_target on vm09 to 3527M 2026-03-10T10:17:49.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:49 vm00 bash[20709]: cluster 2026-03-10T10:17:47.778802+0000 mgr.a (mgr.14150) 149 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:49.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:49 vm00 bash[20709]: cluster 2026-03-10T10:17:47.778802+0000 mgr.a (mgr.14150) 149 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:49.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:49 vm00 bash[20709]: audit 2026-03-10T10:17:48.773335+0000 mon.a (mon.0) 540 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T10:17:49.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:49 vm00 bash[20709]: audit 2026-03-10T10:17:48.773335+0000 mon.a (mon.0) 540 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T10:17:49.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:49 vm00 bash[20709]: cluster 2026-03-10T10:17:48.775840+0000 mon.a (mon.0) 541 : cluster [DBG] osdmap e20: 3 total, 3 up, 3 in 2026-03-10T10:17:49.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:49 vm00 bash[20709]: cluster 2026-03-10T10:17:48.775840+0000 mon.a (mon.0) 541 : cluster [DBG] osdmap e20: 3 total, 3 up, 3 in 2026-03-10T10:17:49.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:49 vm00 bash[20709]: audit 2026-03-10T10:17:48.776664+0000 mon.a (mon.0) 542 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:17:49.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:49 vm00 bash[20709]: audit 2026-03-10T10:17:48.776664+0000 mon.a (mon.0) 542 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:17:51.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:49.776378+0000 mon.a (mon.0) 543 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:49.776378+0000 mon.a (mon.0) 543 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: cluster 2026-03-10T10:17:49.778789+0000 mon.a (mon.0) 544 : cluster [DBG] osdmap e21: 3 total, 3 up, 3 in 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: cluster 2026-03-10T10:17:49.778789+0000 mon.a (mon.0) 544 : cluster [DBG] osdmap e21: 3 total, 3 up, 3 in 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: cluster 2026-03-10T10:17:49.778997+0000 mgr.a (mgr.14150) 150 : cluster [DBG] pgmap v93: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: cluster 2026-03-10T10:17:49.778997+0000 mgr.a (mgr.14150) 150 : cluster [DBG] pgmap v93: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.219652+0000 mon.a (mon.0) 545 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.219652+0000 mon.a (mon.0) 545 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.236715+0000 mon.a (mon.0) 546 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.236715+0000 mon.a (mon.0) 546 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.236983+0000 mon.a (mon.0) 547 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.236983+0000 mon.a (mon.0) 547 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.237042+0000 mon.a (mon.0) 548 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.237042+0000 mon.a (mon.0) 548 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.237350+0000 mon.a (mon.0) 549 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.237350+0000 mon.a (mon.0) 549 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.238179+0000 mon.c (mon.1) 12 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.238179+0000 mon.c (mon.1) 12 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.238695+0000 mon.a (mon.0) 550 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.238695+0000 mon.a (mon.0) 550 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.238740+0000 mon.a (mon.0) 551 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.238740+0000 mon.a (mon.0) 551 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.238987+0000 mon.a (mon.0) 552 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.238987+0000 mon.a (mon.0) 552 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.257161+0000 mon.c (mon.1) 13 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.257161+0000 mon.c (mon.1) 13 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.259043+0000 mon.b (mon.2) 6 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.259043+0000 mon.b (mon.2) 6 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.259323+0000 mon.a (mon.0) 553 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.259323+0000 mon.a (mon.0) 553 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.259408+0000 mon.a (mon.0) 554 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.259408+0000 mon.a (mon.0) 554 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.259500+0000 mon.a (mon.0) 555 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.259500+0000 mon.a (mon.0) 555 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.278181+0000 mon.b (mon.2) 7 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:51.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:50 vm03 bash[21274]: audit 2026-03-10T10:17:50.278181+0000 mon.b (mon.2) 7 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:51.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:49.776378+0000 mon.a (mon.0) 543 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T10:17:51.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:49.776378+0000 mon.a (mon.0) 543 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T10:17:51.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: cluster 2026-03-10T10:17:49.778789+0000 mon.a (mon.0) 544 : cluster [DBG] osdmap e21: 3 total, 3 up, 3 in 2026-03-10T10:17:51.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: cluster 2026-03-10T10:17:49.778789+0000 mon.a (mon.0) 544 : cluster [DBG] osdmap e21: 3 total, 3 up, 3 in 2026-03-10T10:17:51.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: cluster 2026-03-10T10:17:49.778997+0000 mgr.a (mgr.14150) 150 : cluster [DBG] pgmap v93: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:51.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: cluster 2026-03-10T10:17:49.778997+0000 mgr.a (mgr.14150) 150 : cluster [DBG] pgmap v93: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:51.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.219652+0000 mon.a (mon.0) 545 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.219652+0000 mon.a (mon.0) 545 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.236715+0000 mon.a (mon.0) 546 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:51.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.236715+0000 mon.a (mon.0) 546 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:51.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.236983+0000 mon.a (mon.0) 547 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.236983+0000 mon.a (mon.0) 547 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.237042+0000 mon.a (mon.0) 548 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.237042+0000 mon.a (mon.0) 548 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.237350+0000 mon.a (mon.0) 549 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.237350+0000 mon.a (mon.0) 549 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.238179+0000 mon.c (mon.1) 12 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.238179+0000 mon.c (mon.1) 12 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.238695+0000 mon.a (mon.0) 550 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.238695+0000 mon.a (mon.0) 550 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.238740+0000 mon.a (mon.0) 551 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.238740+0000 mon.a (mon.0) 551 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.238987+0000 mon.a (mon.0) 552 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.238987+0000 mon.a (mon.0) 552 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.257161+0000 mon.c (mon.1) 13 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.257161+0000 mon.c (mon.1) 13 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.259043+0000 mon.b (mon.2) 6 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.259043+0000 mon.b (mon.2) 6 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.259323+0000 mon.a (mon.0) 553 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.259323+0000 mon.a (mon.0) 553 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.259408+0000 mon.a (mon.0) 554 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.259408+0000 mon.a (mon.0) 554 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.259500+0000 mon.a (mon.0) 555 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.259500+0000 mon.a (mon.0) 555 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.278181+0000 mon.b (mon.2) 7 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:51.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:50 vm00 bash[20709]: audit 2026-03-10T10:17:50.278181+0000 mon.b (mon.2) 7 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:49.776378+0000 mon.a (mon.0) 543 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:49.776378+0000 mon.a (mon.0) 543 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: cluster 2026-03-10T10:17:49.778789+0000 mon.a (mon.0) 544 : cluster [DBG] osdmap e21: 3 total, 3 up, 3 in 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: cluster 2026-03-10T10:17:49.778789+0000 mon.a (mon.0) 544 : cluster [DBG] osdmap e21: 3 total, 3 up, 3 in 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: cluster 2026-03-10T10:17:49.778997+0000 mgr.a (mgr.14150) 150 : cluster [DBG] pgmap v93: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: cluster 2026-03-10T10:17:49.778997+0000 mgr.a (mgr.14150) 150 : cluster [DBG] pgmap v93: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.219652+0000 mon.a (mon.0) 545 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.219652+0000 mon.a (mon.0) 545 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.236715+0000 mon.a (mon.0) 546 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.236715+0000 mon.a (mon.0) 546 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.236983+0000 mon.a (mon.0) 547 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.236983+0000 mon.a (mon.0) 547 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.237042+0000 mon.a (mon.0) 548 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.237042+0000 mon.a (mon.0) 548 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.237350+0000 mon.a (mon.0) 549 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.237350+0000 mon.a (mon.0) 549 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.238179+0000 mon.c (mon.1) 12 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.238179+0000 mon.c (mon.1) 12 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.238695+0000 mon.a (mon.0) 550 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.238695+0000 mon.a (mon.0) 550 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.238740+0000 mon.a (mon.0) 551 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.238740+0000 mon.a (mon.0) 551 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.238987+0000 mon.a (mon.0) 552 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.238987+0000 mon.a (mon.0) 552 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.257161+0000 mon.c (mon.1) 13 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.257161+0000 mon.c (mon.1) 13 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.259043+0000 mon.b (mon.2) 6 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.259043+0000 mon.b (mon.2) 6 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.259323+0000 mon.a (mon.0) 553 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.259323+0000 mon.a (mon.0) 553 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.259408+0000 mon.a (mon.0) 554 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.259408+0000 mon.a (mon.0) 554 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.259500+0000 mon.a (mon.0) 555 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.259500+0000 mon.a (mon.0) 555 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.278181+0000 mon.b (mon.2) 7 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:51.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:50 vm09 bash[20916]: audit 2026-03-10T10:17:50.278181+0000 mon.b (mon.2) 7 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:17:52.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:51 vm03 bash[21274]: cluster 2026-03-10T10:17:50.791604+0000 mon.a (mon.0) 556 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:17:52.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:51 vm03 bash[21274]: cluster 2026-03-10T10:17:50.791604+0000 mon.a (mon.0) 556 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:17:52.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:51 vm00 bash[20709]: cluster 2026-03-10T10:17:50.791604+0000 mon.a (mon.0) 556 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:17:52.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:51 vm00 bash[20709]: cluster 2026-03-10T10:17:50.791604+0000 mon.a (mon.0) 556 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:17:52.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:51 vm09 bash[20916]: cluster 2026-03-10T10:17:50.791604+0000 mon.a (mon.0) 556 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:17:52.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:51 vm09 bash[20916]: cluster 2026-03-10T10:17:50.791604+0000 mon.a (mon.0) 556 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:17:52.585 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:17:53.100 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:17:53.179 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":22,"num_osds":3,"num_up_osds":3,"osd_up_since":1773137866,"num_in_osds":3,"osd_in_since":1773137850,"num_remapped_pgs":0} 2026-03-10T10:17:53.179 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph osd dump --format=json 2026-03-10T10:17:53.243 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:53 vm03 bash[21274]: cluster 2026-03-10T10:17:51.779196+0000 mgr.a (mgr.14150) 151 : cluster [DBG] pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:53.243 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:53 vm03 bash[21274]: cluster 2026-03-10T10:17:51.779196+0000 mgr.a (mgr.14150) 151 : cluster [DBG] pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:53.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:53 vm00 bash[20709]: cluster 2026-03-10T10:17:51.779196+0000 mgr.a (mgr.14150) 151 : cluster [DBG] pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:53.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:53 vm00 bash[20709]: cluster 2026-03-10T10:17:51.779196+0000 mgr.a (mgr.14150) 151 : cluster [DBG] pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:53.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:53 vm09 bash[20916]: cluster 2026-03-10T10:17:51.779196+0000 mgr.a (mgr.14150) 151 : cluster [DBG] pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:53.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:53 vm09 bash[20916]: cluster 2026-03-10T10:17:51.779196+0000 mgr.a (mgr.14150) 151 : cluster [DBG] pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:54.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:54 vm00 bash[20709]: cluster 2026-03-10T10:17:53.026542+0000 mon.a (mon.0) 557 : cluster [DBG] mgrmap e14: a(active, since 2m), standbys: b 2026-03-10T10:17:54.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:54 vm00 bash[20709]: cluster 2026-03-10T10:17:53.026542+0000 mon.a (mon.0) 557 : cluster [DBG] mgrmap e14: a(active, since 2m), standbys: b 2026-03-10T10:17:54.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:54 vm00 bash[20709]: audit 2026-03-10T10:17:53.100212+0000 mon.a (mon.0) 558 : audit [DBG] from='client.? 192.168.123.100:0/3751830628' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:17:54.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:54 vm00 bash[20709]: audit 2026-03-10T10:17:53.100212+0000 mon.a (mon.0) 558 : audit [DBG] from='client.? 192.168.123.100:0/3751830628' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:17:54.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:54 vm09 bash[20916]: cluster 2026-03-10T10:17:53.026542+0000 mon.a (mon.0) 557 : cluster [DBG] mgrmap e14: a(active, since 2m), standbys: b 2026-03-10T10:17:54.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:54 vm09 bash[20916]: cluster 2026-03-10T10:17:53.026542+0000 mon.a (mon.0) 557 : cluster [DBG] mgrmap e14: a(active, since 2m), standbys: b 2026-03-10T10:17:54.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:54 vm09 bash[20916]: audit 2026-03-10T10:17:53.100212+0000 mon.a (mon.0) 558 : audit [DBG] from='client.? 192.168.123.100:0/3751830628' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:17:54.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:54 vm09 bash[20916]: audit 2026-03-10T10:17:53.100212+0000 mon.a (mon.0) 558 : audit [DBG] from='client.? 192.168.123.100:0/3751830628' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:17:54.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:54 vm03 bash[21274]: cluster 2026-03-10T10:17:53.026542+0000 mon.a (mon.0) 557 : cluster [DBG] mgrmap e14: a(active, since 2m), standbys: b 2026-03-10T10:17:54.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:54 vm03 bash[21274]: cluster 2026-03-10T10:17:53.026542+0000 mon.a (mon.0) 557 : cluster [DBG] mgrmap e14: a(active, since 2m), standbys: b 2026-03-10T10:17:54.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:54 vm03 bash[21274]: audit 2026-03-10T10:17:53.100212+0000 mon.a (mon.0) 558 : audit [DBG] from='client.? 192.168.123.100:0/3751830628' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:17:54.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:54 vm03 bash[21274]: audit 2026-03-10T10:17:53.100212+0000 mon.a (mon.0) 558 : audit [DBG] from='client.? 192.168.123.100:0/3751830628' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:17:55.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:55 vm00 bash[20709]: cluster 2026-03-10T10:17:53.779507+0000 mgr.a (mgr.14150) 152 : cluster [DBG] pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:55.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:55 vm00 bash[20709]: cluster 2026-03-10T10:17:53.779507+0000 mgr.a (mgr.14150) 152 : cluster [DBG] pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:55.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:55 vm09 bash[20916]: cluster 2026-03-10T10:17:53.779507+0000 mgr.a (mgr.14150) 152 : cluster [DBG] pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:55.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:55 vm09 bash[20916]: cluster 2026-03-10T10:17:53.779507+0000 mgr.a (mgr.14150) 152 : cluster [DBG] pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:55.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:55 vm03 bash[21274]: cluster 2026-03-10T10:17:53.779507+0000 mgr.a (mgr.14150) 152 : cluster [DBG] pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:55.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:55 vm03 bash[21274]: cluster 2026-03-10T10:17:53.779507+0000 mgr.a (mgr.14150) 152 : cluster [DBG] pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:57.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:57 vm00 bash[20709]: cluster 2026-03-10T10:17:55.779714+0000 mgr.a (mgr.14150) 153 : cluster [DBG] pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:57.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:57 vm00 bash[20709]: cluster 2026-03-10T10:17:55.779714+0000 mgr.a (mgr.14150) 153 : cluster [DBG] pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:57.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:57 vm09 bash[20916]: cluster 2026-03-10T10:17:55.779714+0000 mgr.a (mgr.14150) 153 : cluster [DBG] pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:57.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:57 vm09 bash[20916]: cluster 2026-03-10T10:17:55.779714+0000 mgr.a (mgr.14150) 153 : cluster [DBG] pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:57.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:57 vm03 bash[21274]: cluster 2026-03-10T10:17:55.779714+0000 mgr.a (mgr.14150) 153 : cluster [DBG] pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:57.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:57 vm03 bash[21274]: cluster 2026-03-10T10:17:55.779714+0000 mgr.a (mgr.14150) 153 : cluster [DBG] pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:57.804 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:17:58.081 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:17:58.081 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":22,"fsid":"9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad","created":"2026-03-10T10:14:45.389041+0000","modified":"2026-03-10T10:17:50.776977+0000","last_up_change":"2026-03-10T10:17:46.426620+0000","last_in_change":"2026-03-10T10:17:30.592213+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T10:17:47.844670+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"22","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":3,"score_stable":3,"optimal_score":1,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"8a37c1d2-927c-46cd-a2b4-e6451084fc61","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6802","nonce":2417190291},{"type":"v1","addr":"192.168.123.100:6803","nonce":2417190291}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6804","nonce":2417190291},{"type":"v1","addr":"192.168.123.100:6805","nonce":2417190291}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6808","nonce":2417190291},{"type":"v1","addr":"192.168.123.100:6809","nonce":2417190291}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6806","nonce":2417190291},{"type":"v1","addr":"192.168.123.100:6807","nonce":2417190291}]},"public_addr":"192.168.123.100:6803/2417190291","cluster_addr":"192.168.123.100:6805/2417190291","heartbeat_back_addr":"192.168.123.100:6809/2417190291","heartbeat_front_addr":"192.168.123.100:6807/2417190291","state":["exists","up"]},{"osd":1,"uuid":"48ce338d-486c-428f-88f7-6f089e65a594","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":20,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6800","nonce":4141501220},{"type":"v1","addr":"192.168.123.103:6801","nonce":4141501220}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6802","nonce":4141501220},{"type":"v1","addr":"192.168.123.103:6803","nonce":4141501220}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6806","nonce":4141501220},{"type":"v1","addr":"192.168.123.103:6807","nonce":4141501220}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6804","nonce":4141501220},{"type":"v1","addr":"192.168.123.103:6805","nonce":4141501220}]},"public_addr":"192.168.123.103:6801/4141501220","cluster_addr":"192.168.123.103:6803/4141501220","heartbeat_back_addr":"192.168.123.103:6807/4141501220","heartbeat_front_addr":"192.168.123.103:6805/4141501220","state":["exists","up"]},{"osd":2,"uuid":"c1f82793-95a2-4beb-bb79-141603b483df","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":18,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":2811766425},{"type":"v1","addr":"192.168.123.109:6801","nonce":2811766425}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":2811766425},{"type":"v1","addr":"192.168.123.109:6803","nonce":2811766425}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":2811766425},{"type":"v1","addr":"192.168.123.109:6807","nonce":2811766425}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":2811766425},{"type":"v1","addr":"192.168.123.109:6805","nonce":2811766425}]},"public_addr":"192.168.123.109:6801/2811766425","cluster_addr":"192.168.123.109:6803/2811766425","heartbeat_back_addr":"192.168.123.109:6807/2811766425","heartbeat_front_addr":"192.168.123.109:6805/2811766425","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:16:41.111365+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:17:14.101290+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:17:45.001792+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.100:6801/521837231":"2026-03-11T10:15:05.753536+0000","192.168.123.100:0/2092203":"2026-03-11T10:15:05.753536+0000","192.168.123.100:0/2293961577":"2026-03-11T10:15:05.753536+0000","192.168.123.100:0/1311426262":"2026-03-11T10:14:55.665189+0000","192.168.123.100:0/3326429106":"2026-03-11T10:14:55.665189+0000","192.168.123.100:0/2087321311":"2026-03-11T10:15:05.753536+0000","192.168.123.100:6800/521837231":"2026-03-11T10:15:05.753536+0000","192.168.123.100:0/1655622803":"2026-03-11T10:14:55.665189+0000","192.168.123.100:6801/3097082882":"2026-03-11T10:14:55.665189+0000","192.168.123.100:6800/3097082882":"2026-03-11T10:14:55.665189+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T10:17:58.134 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-10T10:17:47.844670+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'is_stretch_pool': False, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '22', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}, 'read_balance': {'score_type': 'Fair distribution', 'score_acting': 3, 'score_stable': 3, 'optimal_score': 1, 'raw_score_acting': 3, 'raw_score_stable': 3, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}}] 2026-03-10T10:17:58.134 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph osd pool get .mgr pg_num 2026-03-10T10:17:58.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:58 vm00 bash[20709]: audit 2026-03-10T10:17:57.333578+0000 mon.a (mon.0) 559 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:58 vm00 bash[20709]: audit 2026-03-10T10:17:57.333578+0000 mon.a (mon.0) 559 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:58 vm00 bash[20709]: audit 2026-03-10T10:17:57.338195+0000 mon.a (mon.0) 560 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:58 vm00 bash[20709]: audit 2026-03-10T10:17:57.338195+0000 mon.a (mon.0) 560 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:58 vm00 bash[20709]: audit 2026-03-10T10:17:57.341990+0000 mon.a (mon.0) 561 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:58 vm00 bash[20709]: audit 2026-03-10T10:17:57.341990+0000 mon.a (mon.0) 561 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:58 vm00 bash[20709]: audit 2026-03-10T10:17:57.873822+0000 mon.a (mon.0) 562 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:58 vm00 bash[20709]: audit 2026-03-10T10:17:57.873822+0000 mon.a (mon.0) 562 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:58 vm00 bash[20709]: audit 2026-03-10T10:17:57.877885+0000 mon.a (mon.0) 563 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:58 vm00 bash[20709]: audit 2026-03-10T10:17:57.877885+0000 mon.a (mon.0) 563 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:58 vm00 bash[20709]: audit 2026-03-10T10:17:57.881871+0000 mon.a (mon.0) 564 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:58 vm00 bash[20709]: audit 2026-03-10T10:17:57.881871+0000 mon.a (mon.0) 564 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:58 vm00 bash[20709]: audit 2026-03-10T10:17:58.081346+0000 mon.a (mon.0) 565 : audit [DBG] from='client.? 192.168.123.100:0/2192361101' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:17:58.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:58 vm00 bash[20709]: audit 2026-03-10T10:17:58.081346+0000 mon.a (mon.0) 565 : audit [DBG] from='client.? 192.168.123.100:0/2192361101' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:17:58.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:58 vm09 bash[20916]: audit 2026-03-10T10:17:57.333578+0000 mon.a (mon.0) 559 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:58 vm09 bash[20916]: audit 2026-03-10T10:17:57.333578+0000 mon.a (mon.0) 559 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:58 vm09 bash[20916]: audit 2026-03-10T10:17:57.338195+0000 mon.a (mon.0) 560 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:58 vm09 bash[20916]: audit 2026-03-10T10:17:57.338195+0000 mon.a (mon.0) 560 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:58 vm09 bash[20916]: audit 2026-03-10T10:17:57.341990+0000 mon.a (mon.0) 561 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:58 vm09 bash[20916]: audit 2026-03-10T10:17:57.341990+0000 mon.a (mon.0) 561 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:58 vm09 bash[20916]: audit 2026-03-10T10:17:57.873822+0000 mon.a (mon.0) 562 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:58 vm09 bash[20916]: audit 2026-03-10T10:17:57.873822+0000 mon.a (mon.0) 562 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:58 vm09 bash[20916]: audit 2026-03-10T10:17:57.877885+0000 mon.a (mon.0) 563 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:58 vm09 bash[20916]: audit 2026-03-10T10:17:57.877885+0000 mon.a (mon.0) 563 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:58 vm09 bash[20916]: audit 2026-03-10T10:17:57.881871+0000 mon.a (mon.0) 564 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:58 vm09 bash[20916]: audit 2026-03-10T10:17:57.881871+0000 mon.a (mon.0) 564 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:58 vm09 bash[20916]: audit 2026-03-10T10:17:58.081346+0000 mon.a (mon.0) 565 : audit [DBG] from='client.? 192.168.123.100:0/2192361101' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:17:58.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:58 vm09 bash[20916]: audit 2026-03-10T10:17:58.081346+0000 mon.a (mon.0) 565 : audit [DBG] from='client.? 192.168.123.100:0/2192361101' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:17:58.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:58 vm03 bash[21274]: audit 2026-03-10T10:17:57.333578+0000 mon.a (mon.0) 559 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:58 vm03 bash[21274]: audit 2026-03-10T10:17:57.333578+0000 mon.a (mon.0) 559 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:58 vm03 bash[21274]: audit 2026-03-10T10:17:57.338195+0000 mon.a (mon.0) 560 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:58 vm03 bash[21274]: audit 2026-03-10T10:17:57.338195+0000 mon.a (mon.0) 560 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:58 vm03 bash[21274]: audit 2026-03-10T10:17:57.341990+0000 mon.a (mon.0) 561 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:58 vm03 bash[21274]: audit 2026-03-10T10:17:57.341990+0000 mon.a (mon.0) 561 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:58 vm03 bash[21274]: audit 2026-03-10T10:17:57.873822+0000 mon.a (mon.0) 562 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:58 vm03 bash[21274]: audit 2026-03-10T10:17:57.873822+0000 mon.a (mon.0) 562 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:58 vm03 bash[21274]: audit 2026-03-10T10:17:57.877885+0000 mon.a (mon.0) 563 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:58 vm03 bash[21274]: audit 2026-03-10T10:17:57.877885+0000 mon.a (mon.0) 563 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:58 vm03 bash[21274]: audit 2026-03-10T10:17:57.881871+0000 mon.a (mon.0) 564 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:58 vm03 bash[21274]: audit 2026-03-10T10:17:57.881871+0000 mon.a (mon.0) 564 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:58.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:58 vm03 bash[21274]: audit 2026-03-10T10:17:58.081346+0000 mon.a (mon.0) 565 : audit [DBG] from='client.? 192.168.123.100:0/2192361101' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:17:58.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:58 vm03 bash[21274]: audit 2026-03-10T10:17:58.081346+0000 mon.a (mon.0) 565 : audit [DBG] from='client.? 192.168.123.100:0/2192361101' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:17:59.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:59 vm00 bash[20709]: cluster 2026-03-10T10:17:57.779990+0000 mgr.a (mgr.14150) 154 : cluster [DBG] pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:59.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:59 vm00 bash[20709]: cluster 2026-03-10T10:17:57.779990+0000 mgr.a (mgr.14150) 154 : cluster [DBG] pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:59.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:59 vm00 bash[20709]: audit 2026-03-10T10:17:59.333992+0000 mon.a (mon.0) 566 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:59 vm00 bash[20709]: audit 2026-03-10T10:17:59.333992+0000 mon.a (mon.0) 566 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:59 vm00 bash[20709]: audit 2026-03-10T10:17:59.339186+0000 mon.a (mon.0) 567 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:59 vm00 bash[20709]: audit 2026-03-10T10:17:59.339186+0000 mon.a (mon.0) 567 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:59 vm00 bash[20709]: audit 2026-03-10T10:17:59.340319+0000 mon.a (mon.0) 568 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:59.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:59 vm00 bash[20709]: audit 2026-03-10T10:17:59.340319+0000 mon.a (mon.0) 568 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:59.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:59 vm00 bash[20709]: audit 2026-03-10T10:17:59.342572+0000 mon.a (mon.0) 569 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:59.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:59 vm00 bash[20709]: audit 2026-03-10T10:17:59.342572+0000 mon.a (mon.0) 569 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:59.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:59 vm00 bash[20709]: audit 2026-03-10T10:17:59.344505+0000 mon.a (mon.0) 570 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:59 vm00 bash[20709]: audit 2026-03-10T10:17:59.344505+0000 mon.a (mon.0) 570 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:59 vm00 bash[20709]: audit 2026-03-10T10:17:59.344684+0000 mon.a (mon.0) 571 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:59.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:59 vm00 bash[20709]: audit 2026-03-10T10:17:59.344684+0000 mon.a (mon.0) 571 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:59.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:59 vm00 bash[20709]: audit 2026-03-10T10:17:59.350373+0000 mon.a (mon.0) 572 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:17:59 vm00 bash[20709]: audit 2026-03-10T10:17:59.350373+0000 mon.a (mon.0) 572 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:59 vm09 bash[20916]: cluster 2026-03-10T10:17:57.779990+0000 mgr.a (mgr.14150) 154 : cluster [DBG] pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:59.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:59 vm09 bash[20916]: cluster 2026-03-10T10:17:57.779990+0000 mgr.a (mgr.14150) 154 : cluster [DBG] pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:59.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:59 vm09 bash[20916]: audit 2026-03-10T10:17:59.333992+0000 mon.a (mon.0) 566 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:59 vm09 bash[20916]: audit 2026-03-10T10:17:59.333992+0000 mon.a (mon.0) 566 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:59 vm09 bash[20916]: audit 2026-03-10T10:17:59.339186+0000 mon.a (mon.0) 567 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:59 vm09 bash[20916]: audit 2026-03-10T10:17:59.339186+0000 mon.a (mon.0) 567 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:59 vm09 bash[20916]: audit 2026-03-10T10:17:59.340319+0000 mon.a (mon.0) 568 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:59.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:59 vm09 bash[20916]: audit 2026-03-10T10:17:59.340319+0000 mon.a (mon.0) 568 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:59.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:59 vm09 bash[20916]: audit 2026-03-10T10:17:59.342572+0000 mon.a (mon.0) 569 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:59.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:59 vm09 bash[20916]: audit 2026-03-10T10:17:59.342572+0000 mon.a (mon.0) 569 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:59.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:59 vm09 bash[20916]: audit 2026-03-10T10:17:59.344505+0000 mon.a (mon.0) 570 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:59 vm09 bash[20916]: audit 2026-03-10T10:17:59.344505+0000 mon.a (mon.0) 570 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:59 vm09 bash[20916]: audit 2026-03-10T10:17:59.344684+0000 mon.a (mon.0) 571 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:59.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:59 vm09 bash[20916]: audit 2026-03-10T10:17:59.344684+0000 mon.a (mon.0) 571 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:59.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:59 vm09 bash[20916]: audit 2026-03-10T10:17:59.350373+0000 mon.a (mon.0) 572 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:17:59 vm09 bash[20916]: audit 2026-03-10T10:17:59.350373+0000 mon.a (mon.0) 572 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:59 vm03 bash[21274]: cluster 2026-03-10T10:17:57.779990+0000 mgr.a (mgr.14150) 154 : cluster [DBG] pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:59.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:59 vm03 bash[21274]: cluster 2026-03-10T10:17:57.779990+0000 mgr.a (mgr.14150) 154 : cluster [DBG] pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:17:59.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:59 vm03 bash[21274]: audit 2026-03-10T10:17:59.333992+0000 mon.a (mon.0) 566 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:59 vm03 bash[21274]: audit 2026-03-10T10:17:59.333992+0000 mon.a (mon.0) 566 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:59 vm03 bash[21274]: audit 2026-03-10T10:17:59.339186+0000 mon.a (mon.0) 567 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:59 vm03 bash[21274]: audit 2026-03-10T10:17:59.339186+0000 mon.a (mon.0) 567 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:59 vm03 bash[21274]: audit 2026-03-10T10:17:59.340319+0000 mon.a (mon.0) 568 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:59.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:59 vm03 bash[21274]: audit 2026-03-10T10:17:59.340319+0000 mon.a (mon.0) 568 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:17:59.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:59 vm03 bash[21274]: audit 2026-03-10T10:17:59.342572+0000 mon.a (mon.0) 569 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:59.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:59 vm03 bash[21274]: audit 2026-03-10T10:17:59.342572+0000 mon.a (mon.0) 569 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:17:59.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:59 vm03 bash[21274]: audit 2026-03-10T10:17:59.344505+0000 mon.a (mon.0) 570 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:59 vm03 bash[21274]: audit 2026-03-10T10:17:59.344505+0000 mon.a (mon.0) 570 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:59 vm03 bash[21274]: audit 2026-03-10T10:17:59.344684+0000 mon.a (mon.0) 571 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:59.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:59 vm03 bash[21274]: audit 2026-03-10T10:17:59.344684+0000 mon.a (mon.0) 571 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:17:59.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:59 vm03 bash[21274]: audit 2026-03-10T10:17:59.350373+0000 mon.a (mon.0) 572 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:17:59.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:17:59 vm03 bash[21274]: audit 2026-03-10T10:17:59.350373+0000 mon.a (mon.0) 572 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:00.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:00 vm00 bash[20709]: cephadm 2026-03-10T10:17:59.339824+0000 mgr.a (mgr.14150) 155 : cephadm [INF] Detected new or changed devices on vm09 2026-03-10T10:18:00.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:00 vm00 bash[20709]: cephadm 2026-03-10T10:17:59.339824+0000 mgr.a (mgr.14150) 155 : cephadm [INF] Detected new or changed devices on vm09 2026-03-10T10:18:00.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:00 vm09 bash[20916]: cephadm 2026-03-10T10:17:59.339824+0000 mgr.a (mgr.14150) 155 : cephadm [INF] Detected new or changed devices on vm09 2026-03-10T10:18:00.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:00 vm09 bash[20916]: cephadm 2026-03-10T10:17:59.339824+0000 mgr.a (mgr.14150) 155 : cephadm [INF] Detected new or changed devices on vm09 2026-03-10T10:18:00.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:00 vm03 bash[21274]: cephadm 2026-03-10T10:17:59.339824+0000 mgr.a (mgr.14150) 155 : cephadm [INF] Detected new or changed devices on vm09 2026-03-10T10:18:00.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:00 vm03 bash[21274]: cephadm 2026-03-10T10:17:59.339824+0000 mgr.a (mgr.14150) 155 : cephadm [INF] Detected new or changed devices on vm09 2026-03-10T10:18:01.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:01 vm00 bash[20709]: cluster 2026-03-10T10:17:59.780231+0000 mgr.a (mgr.14150) 156 : cluster [DBG] pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:01.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:01 vm00 bash[20709]: cluster 2026-03-10T10:17:59.780231+0000 mgr.a (mgr.14150) 156 : cluster [DBG] pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:01.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:01 vm09 bash[20916]: cluster 2026-03-10T10:17:59.780231+0000 mgr.a (mgr.14150) 156 : cluster [DBG] pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:01.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:01 vm09 bash[20916]: cluster 2026-03-10T10:17:59.780231+0000 mgr.a (mgr.14150) 156 : cluster [DBG] pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:01.817 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:18:01.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:01 vm03 bash[21274]: cluster 2026-03-10T10:17:59.780231+0000 mgr.a (mgr.14150) 156 : cluster [DBG] pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:01.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:01 vm03 bash[21274]: cluster 2026-03-10T10:17:59.780231+0000 mgr.a (mgr.14150) 156 : cluster [DBG] pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:02.067 INFO:teuthology.orchestra.run.vm00.stdout:pg_num: 1 2026-03-10T10:18:02.123 INFO:tasks.cephadm:Setting up client nodes... 2026-03-10T10:18:02.123 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-10T10:18:02.123 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-10T10:18:02.123 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph mgr dump --format=json 2026-03-10T10:18:02.379 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:02 vm00 bash[20709]: audit 2026-03-10T10:18:02.067523+0000 mon.a (mon.0) 573 : audit [DBG] from='client.? 192.168.123.100:0/3452107197' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T10:18:02.379 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:02 vm00 bash[20709]: audit 2026-03-10T10:18:02.067523+0000 mon.a (mon.0) 573 : audit [DBG] from='client.? 192.168.123.100:0/3452107197' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T10:18:02.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:02 vm09 bash[20916]: audit 2026-03-10T10:18:02.067523+0000 mon.a (mon.0) 573 : audit [DBG] from='client.? 192.168.123.100:0/3452107197' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T10:18:02.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:02 vm09 bash[20916]: audit 2026-03-10T10:18:02.067523+0000 mon.a (mon.0) 573 : audit [DBG] from='client.? 192.168.123.100:0/3452107197' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T10:18:02.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:02 vm03 bash[21274]: audit 2026-03-10T10:18:02.067523+0000 mon.a (mon.0) 573 : audit [DBG] from='client.? 192.168.123.100:0/3452107197' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T10:18:02.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:02 vm03 bash[21274]: audit 2026-03-10T10:18:02.067523+0000 mon.a (mon.0) 573 : audit [DBG] from='client.? 192.168.123.100:0/3452107197' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T10:18:03.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:03 vm00 bash[20709]: cluster 2026-03-10T10:18:01.780462+0000 mgr.a (mgr.14150) 157 : cluster [DBG] pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:03.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:03 vm00 bash[20709]: cluster 2026-03-10T10:18:01.780462+0000 mgr.a (mgr.14150) 157 : cluster [DBG] pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:03.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:03 vm00 bash[20709]: audit 2026-03-10T10:18:02.948886+0000 mon.a (mon.0) 574 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:03 vm00 bash[20709]: audit 2026-03-10T10:18:02.948886+0000 mon.a (mon.0) 574 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:03 vm00 bash[20709]: audit 2026-03-10T10:18:02.953077+0000 mon.a (mon.0) 575 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:03 vm00 bash[20709]: audit 2026-03-10T10:18:02.953077+0000 mon.a (mon.0) 575 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:03 vm00 bash[20709]: audit 2026-03-10T10:18:02.955019+0000 mon.a (mon.0) 576 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:03.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:03 vm00 bash[20709]: audit 2026-03-10T10:18:02.955019+0000 mon.a (mon.0) 576 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:03.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:03 vm00 bash[20709]: audit 2026-03-10T10:18:02.956252+0000 mon.a (mon.0) 577 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:03.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:03 vm00 bash[20709]: audit 2026-03-10T10:18:02.956252+0000 mon.a (mon.0) 577 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:03.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:03 vm00 bash[20709]: audit 2026-03-10T10:18:02.958239+0000 mon.a (mon.0) 578 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:03 vm00 bash[20709]: audit 2026-03-10T10:18:02.958239+0000 mon.a (mon.0) 578 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:03 vm00 bash[20709]: audit 2026-03-10T10:18:02.958903+0000 mon.a (mon.0) 579 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:03.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:03 vm00 bash[20709]: audit 2026-03-10T10:18:02.958903+0000 mon.a (mon.0) 579 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:03.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:03 vm00 bash[20709]: audit 2026-03-10T10:18:02.963163+0000 mon.a (mon.0) 580 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.667 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:03 vm00 bash[20709]: audit 2026-03-10T10:18:02.963163+0000 mon.a (mon.0) 580 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:03 vm09 bash[20916]: cluster 2026-03-10T10:18:01.780462+0000 mgr.a (mgr.14150) 157 : cluster [DBG] pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:03.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:03 vm09 bash[20916]: cluster 2026-03-10T10:18:01.780462+0000 mgr.a (mgr.14150) 157 : cluster [DBG] pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:03.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:03 vm09 bash[20916]: audit 2026-03-10T10:18:02.948886+0000 mon.a (mon.0) 574 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:03 vm09 bash[20916]: audit 2026-03-10T10:18:02.948886+0000 mon.a (mon.0) 574 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:03 vm09 bash[20916]: audit 2026-03-10T10:18:02.953077+0000 mon.a (mon.0) 575 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:03 vm09 bash[20916]: audit 2026-03-10T10:18:02.953077+0000 mon.a (mon.0) 575 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:03 vm09 bash[20916]: audit 2026-03-10T10:18:02.955019+0000 mon.a (mon.0) 576 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:03.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:03 vm09 bash[20916]: audit 2026-03-10T10:18:02.955019+0000 mon.a (mon.0) 576 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:03.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:03 vm09 bash[20916]: audit 2026-03-10T10:18:02.956252+0000 mon.a (mon.0) 577 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:03.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:03 vm09 bash[20916]: audit 2026-03-10T10:18:02.956252+0000 mon.a (mon.0) 577 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:03.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:03 vm09 bash[20916]: audit 2026-03-10T10:18:02.958239+0000 mon.a (mon.0) 578 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:03 vm09 bash[20916]: audit 2026-03-10T10:18:02.958239+0000 mon.a (mon.0) 578 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:03 vm09 bash[20916]: audit 2026-03-10T10:18:02.958903+0000 mon.a (mon.0) 579 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:03.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:03 vm09 bash[20916]: audit 2026-03-10T10:18:02.958903+0000 mon.a (mon.0) 579 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:03.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:03 vm09 bash[20916]: audit 2026-03-10T10:18:02.963163+0000 mon.a (mon.0) 580 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:03 vm09 bash[20916]: audit 2026-03-10T10:18:02.963163+0000 mon.a (mon.0) 580 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:03 vm03 bash[21274]: cluster 2026-03-10T10:18:01.780462+0000 mgr.a (mgr.14150) 157 : cluster [DBG] pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:03.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:03 vm03 bash[21274]: cluster 2026-03-10T10:18:01.780462+0000 mgr.a (mgr.14150) 157 : cluster [DBG] pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:03.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:03 vm03 bash[21274]: audit 2026-03-10T10:18:02.948886+0000 mon.a (mon.0) 574 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:03 vm03 bash[21274]: audit 2026-03-10T10:18:02.948886+0000 mon.a (mon.0) 574 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:03 vm03 bash[21274]: audit 2026-03-10T10:18:02.953077+0000 mon.a (mon.0) 575 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:03 vm03 bash[21274]: audit 2026-03-10T10:18:02.953077+0000 mon.a (mon.0) 575 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:03 vm03 bash[21274]: audit 2026-03-10T10:18:02.955019+0000 mon.a (mon.0) 576 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:03.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:03 vm03 bash[21274]: audit 2026-03-10T10:18:02.955019+0000 mon.a (mon.0) 576 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:03.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:03 vm03 bash[21274]: audit 2026-03-10T10:18:02.956252+0000 mon.a (mon.0) 577 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:03.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:03 vm03 bash[21274]: audit 2026-03-10T10:18:02.956252+0000 mon.a (mon.0) 577 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:03.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:03 vm03 bash[21274]: audit 2026-03-10T10:18:02.958239+0000 mon.a (mon.0) 578 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:03 vm03 bash[21274]: audit 2026-03-10T10:18:02.958239+0000 mon.a (mon.0) 578 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:03 vm03 bash[21274]: audit 2026-03-10T10:18:02.958903+0000 mon.a (mon.0) 579 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:03.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:03 vm03 bash[21274]: audit 2026-03-10T10:18:02.958903+0000 mon.a (mon.0) 579 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:03.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:03 vm03 bash[21274]: audit 2026-03-10T10:18:02.963163+0000 mon.a (mon.0) 580 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:03.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:03 vm03 bash[21274]: audit 2026-03-10T10:18:02.963163+0000 mon.a (mon.0) 580 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:05.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:05 vm09 bash[20916]: cluster 2026-03-10T10:18:03.780727+0000 mgr.a (mgr.14150) 158 : cluster [DBG] pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:05.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:05 vm09 bash[20916]: cluster 2026-03-10T10:18:03.780727+0000 mgr.a (mgr.14150) 158 : cluster [DBG] pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:05.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:05 vm00 bash[20709]: cluster 2026-03-10T10:18:03.780727+0000 mgr.a (mgr.14150) 158 : cluster [DBG] pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:05.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:05 vm00 bash[20709]: cluster 2026-03-10T10:18:03.780727+0000 mgr.a (mgr.14150) 158 : cluster [DBG] pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:05.827 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:18:05.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:05 vm03 bash[21274]: cluster 2026-03-10T10:18:03.780727+0000 mgr.a (mgr.14150) 158 : cluster [DBG] pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:05.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:05 vm03 bash[21274]: cluster 2026-03-10T10:18:03.780727+0000 mgr.a (mgr.14150) 158 : cluster [DBG] pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:06.158 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:18:06.210 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":14,"flags":0,"active_gid":14150,"active_name":"a","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6800","nonce":273418958},{"type":"v1","addr":"192.168.123.100:6801","nonce":273418958}]},"active_addr":"192.168.123.100:6801/273418958","active_change":"2026-03-10T10:15:05.753633+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":24112,"name":"b","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.100:8443/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":3,"active_clients":[{"name":"devicehealth","addrvec":[{"type":"v2","addr":"192.168.123.100:0","nonce":2772701187}]},{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.100:0","nonce":3168737497}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.100:0","nonce":2713171197}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.100:0","nonce":2067118339}]}]} 2026-03-10T10:18:06.212 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-10T10:18:06.212 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-10T10:18:06.212 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph osd dump --format=json 2026-03-10T10:18:06.475 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:06 vm00 bash[20709]: audit 2026-03-10T10:18:06.156875+0000 mon.a (mon.0) 581 : audit [DBG] from='client.? 192.168.123.100:0/1233477620' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T10:18:06.475 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:06 vm00 bash[20709]: audit 2026-03-10T10:18:06.156875+0000 mon.a (mon.0) 581 : audit [DBG] from='client.? 192.168.123.100:0/1233477620' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T10:18:06.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:06 vm09 bash[20916]: audit 2026-03-10T10:18:06.156875+0000 mon.a (mon.0) 581 : audit [DBG] from='client.? 192.168.123.100:0/1233477620' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T10:18:06.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:06 vm09 bash[20916]: audit 2026-03-10T10:18:06.156875+0000 mon.a (mon.0) 581 : audit [DBG] from='client.? 192.168.123.100:0/1233477620' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T10:18:06.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:06 vm03 bash[21274]: audit 2026-03-10T10:18:06.156875+0000 mon.a (mon.0) 581 : audit [DBG] from='client.? 192.168.123.100:0/1233477620' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T10:18:06.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:06 vm03 bash[21274]: audit 2026-03-10T10:18:06.156875+0000 mon.a (mon.0) 581 : audit [DBG] from='client.? 192.168.123.100:0/1233477620' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T10:18:07.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:07 vm09 bash[20916]: cluster 2026-03-10T10:18:05.780935+0000 mgr.a (mgr.14150) 159 : cluster [DBG] pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:07.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:07 vm09 bash[20916]: cluster 2026-03-10T10:18:05.780935+0000 mgr.a (mgr.14150) 159 : cluster [DBG] pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:07.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:07 vm03 bash[21274]: cluster 2026-03-10T10:18:05.780935+0000 mgr.a (mgr.14150) 159 : cluster [DBG] pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:07.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:07 vm03 bash[21274]: cluster 2026-03-10T10:18:05.780935+0000 mgr.a (mgr.14150) 159 : cluster [DBG] pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:07.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:07 vm00 bash[20709]: cluster 2026-03-10T10:18:05.780935+0000 mgr.a (mgr.14150) 159 : cluster [DBG] pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:07.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:07 vm00 bash[20709]: cluster 2026-03-10T10:18:05.780935+0000 mgr.a (mgr.14150) 159 : cluster [DBG] pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:09.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:09 vm00 bash[20709]: cluster 2026-03-10T10:18:07.781238+0000 mgr.a (mgr.14150) 160 : cluster [DBG] pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:09.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:09 vm00 bash[20709]: cluster 2026-03-10T10:18:07.781238+0000 mgr.a (mgr.14150) 160 : cluster [DBG] pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:10.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:09 vm09 bash[20916]: cluster 2026-03-10T10:18:07.781238+0000 mgr.a (mgr.14150) 160 : cluster [DBG] pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:10.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:09 vm09 bash[20916]: cluster 2026-03-10T10:18:07.781238+0000 mgr.a (mgr.14150) 160 : cluster [DBG] pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:10.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:09 vm03 bash[21274]: cluster 2026-03-10T10:18:07.781238+0000 mgr.a (mgr.14150) 160 : cluster [DBG] pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:10.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:09 vm03 bash[21274]: cluster 2026-03-10T10:18:07.781238+0000 mgr.a (mgr.14150) 160 : cluster [DBG] pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:10.207 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:18:10.523 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:18:10.523 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":22,"fsid":"9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad","created":"2026-03-10T10:14:45.389041+0000","modified":"2026-03-10T10:17:50.776977+0000","last_up_change":"2026-03-10T10:17:46.426620+0000","last_in_change":"2026-03-10T10:17:30.592213+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T10:17:47.844670+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"22","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":3,"score_stable":3,"optimal_score":1,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"8a37c1d2-927c-46cd-a2b4-e6451084fc61","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6802","nonce":2417190291},{"type":"v1","addr":"192.168.123.100:6803","nonce":2417190291}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6804","nonce":2417190291},{"type":"v1","addr":"192.168.123.100:6805","nonce":2417190291}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6808","nonce":2417190291},{"type":"v1","addr":"192.168.123.100:6809","nonce":2417190291}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6806","nonce":2417190291},{"type":"v1","addr":"192.168.123.100:6807","nonce":2417190291}]},"public_addr":"192.168.123.100:6803/2417190291","cluster_addr":"192.168.123.100:6805/2417190291","heartbeat_back_addr":"192.168.123.100:6809/2417190291","heartbeat_front_addr":"192.168.123.100:6807/2417190291","state":["exists","up"]},{"osd":1,"uuid":"48ce338d-486c-428f-88f7-6f089e65a594","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":20,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6800","nonce":4141501220},{"type":"v1","addr":"192.168.123.103:6801","nonce":4141501220}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6802","nonce":4141501220},{"type":"v1","addr":"192.168.123.103:6803","nonce":4141501220}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6806","nonce":4141501220},{"type":"v1","addr":"192.168.123.103:6807","nonce":4141501220}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6804","nonce":4141501220},{"type":"v1","addr":"192.168.123.103:6805","nonce":4141501220}]},"public_addr":"192.168.123.103:6801/4141501220","cluster_addr":"192.168.123.103:6803/4141501220","heartbeat_back_addr":"192.168.123.103:6807/4141501220","heartbeat_front_addr":"192.168.123.103:6805/4141501220","state":["exists","up"]},{"osd":2,"uuid":"c1f82793-95a2-4beb-bb79-141603b483df","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":18,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":2811766425},{"type":"v1","addr":"192.168.123.109:6801","nonce":2811766425}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":2811766425},{"type":"v1","addr":"192.168.123.109:6803","nonce":2811766425}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":2811766425},{"type":"v1","addr":"192.168.123.109:6807","nonce":2811766425}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":2811766425},{"type":"v1","addr":"192.168.123.109:6805","nonce":2811766425}]},"public_addr":"192.168.123.109:6801/2811766425","cluster_addr":"192.168.123.109:6803/2811766425","heartbeat_back_addr":"192.168.123.109:6807/2811766425","heartbeat_front_addr":"192.168.123.109:6805/2811766425","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:16:41.111365+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:17:14.101290+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:17:45.001792+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.100:6801/521837231":"2026-03-11T10:15:05.753536+0000","192.168.123.100:0/2092203":"2026-03-11T10:15:05.753536+0000","192.168.123.100:0/2293961577":"2026-03-11T10:15:05.753536+0000","192.168.123.100:0/1311426262":"2026-03-11T10:14:55.665189+0000","192.168.123.100:0/3326429106":"2026-03-11T10:14:55.665189+0000","192.168.123.100:0/2087321311":"2026-03-11T10:15:05.753536+0000","192.168.123.100:6800/521837231":"2026-03-11T10:15:05.753536+0000","192.168.123.100:0/1655622803":"2026-03-11T10:14:55.665189+0000","192.168.123.100:6801/3097082882":"2026-03-11T10:14:55.665189+0000","192.168.123.100:6800/3097082882":"2026-03-11T10:14:55.665189+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T10:18:10.585 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-10T10:18:10.585 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph osd dump --format=json 2026-03-10T10:18:10.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:10 vm00 bash[20709]: audit 2026-03-10T10:18:10.523423+0000 mon.a (mon.0) 582 : audit [DBG] from='client.? 192.168.123.100:0/382139207' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:18:10.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:10 vm00 bash[20709]: audit 2026-03-10T10:18:10.523423+0000 mon.a (mon.0) 582 : audit [DBG] from='client.? 192.168.123.100:0/382139207' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:18:11.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:10 vm09 bash[20916]: audit 2026-03-10T10:18:10.523423+0000 mon.a (mon.0) 582 : audit [DBG] from='client.? 192.168.123.100:0/382139207' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:18:11.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:10 vm09 bash[20916]: audit 2026-03-10T10:18:10.523423+0000 mon.a (mon.0) 582 : audit [DBG] from='client.? 192.168.123.100:0/382139207' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:18:11.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:10 vm03 bash[21274]: audit 2026-03-10T10:18:10.523423+0000 mon.a (mon.0) 582 : audit [DBG] from='client.? 192.168.123.100:0/382139207' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:18:11.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:10 vm03 bash[21274]: audit 2026-03-10T10:18:10.523423+0000 mon.a (mon.0) 582 : audit [DBG] from='client.? 192.168.123.100:0/382139207' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:18:11.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:11 vm00 bash[20709]: cluster 2026-03-10T10:18:09.781550+0000 mgr.a (mgr.14150) 161 : cluster [DBG] pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:11.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:11 vm00 bash[20709]: cluster 2026-03-10T10:18:09.781550+0000 mgr.a (mgr.14150) 161 : cluster [DBG] pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:12.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:11 vm09 bash[20916]: cluster 2026-03-10T10:18:09.781550+0000 mgr.a (mgr.14150) 161 : cluster [DBG] pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:12.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:11 vm09 bash[20916]: cluster 2026-03-10T10:18:09.781550+0000 mgr.a (mgr.14150) 161 : cluster [DBG] pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:12.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:11 vm03 bash[21274]: cluster 2026-03-10T10:18:09.781550+0000 mgr.a (mgr.14150) 161 : cluster [DBG] pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:12.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:11 vm03 bash[21274]: cluster 2026-03-10T10:18:09.781550+0000 mgr.a (mgr.14150) 161 : cluster [DBG] pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:13.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:13 vm00 bash[20709]: cluster 2026-03-10T10:18:11.781794+0000 mgr.a (mgr.14150) 162 : cluster [DBG] pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:13.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:13 vm00 bash[20709]: cluster 2026-03-10T10:18:11.781794+0000 mgr.a (mgr.14150) 162 : cluster [DBG] pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:14.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:13 vm09 bash[20916]: cluster 2026-03-10T10:18:11.781794+0000 mgr.a (mgr.14150) 162 : cluster [DBG] pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:14.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:13 vm09 bash[20916]: cluster 2026-03-10T10:18:11.781794+0000 mgr.a (mgr.14150) 162 : cluster [DBG] pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:14.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:13 vm03 bash[21274]: cluster 2026-03-10T10:18:11.781794+0000 mgr.a (mgr.14150) 162 : cluster [DBG] pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:14.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:13 vm03 bash[21274]: cluster 2026-03-10T10:18:11.781794+0000 mgr.a (mgr.14150) 162 : cluster [DBG] pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:15.206 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:18:15.494 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:18:15.494 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":22,"fsid":"9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad","created":"2026-03-10T10:14:45.389041+0000","modified":"2026-03-10T10:17:50.776977+0000","last_up_change":"2026-03-10T10:17:46.426620+0000","last_in_change":"2026-03-10T10:17:30.592213+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T10:17:47.844670+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"22","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":3,"score_stable":3,"optimal_score":1,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"8a37c1d2-927c-46cd-a2b4-e6451084fc61","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6802","nonce":2417190291},{"type":"v1","addr":"192.168.123.100:6803","nonce":2417190291}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6804","nonce":2417190291},{"type":"v1","addr":"192.168.123.100:6805","nonce":2417190291}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6808","nonce":2417190291},{"type":"v1","addr":"192.168.123.100:6809","nonce":2417190291}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6806","nonce":2417190291},{"type":"v1","addr":"192.168.123.100:6807","nonce":2417190291}]},"public_addr":"192.168.123.100:6803/2417190291","cluster_addr":"192.168.123.100:6805/2417190291","heartbeat_back_addr":"192.168.123.100:6809/2417190291","heartbeat_front_addr":"192.168.123.100:6807/2417190291","state":["exists","up"]},{"osd":1,"uuid":"48ce338d-486c-428f-88f7-6f089e65a594","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":20,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6800","nonce":4141501220},{"type":"v1","addr":"192.168.123.103:6801","nonce":4141501220}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6802","nonce":4141501220},{"type":"v1","addr":"192.168.123.103:6803","nonce":4141501220}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6806","nonce":4141501220},{"type":"v1","addr":"192.168.123.103:6807","nonce":4141501220}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6804","nonce":4141501220},{"type":"v1","addr":"192.168.123.103:6805","nonce":4141501220}]},"public_addr":"192.168.123.103:6801/4141501220","cluster_addr":"192.168.123.103:6803/4141501220","heartbeat_back_addr":"192.168.123.103:6807/4141501220","heartbeat_front_addr":"192.168.123.103:6805/4141501220","state":["exists","up"]},{"osd":2,"uuid":"c1f82793-95a2-4beb-bb79-141603b483df","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":18,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":2811766425},{"type":"v1","addr":"192.168.123.109:6801","nonce":2811766425}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":2811766425},{"type":"v1","addr":"192.168.123.109:6803","nonce":2811766425}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":2811766425},{"type":"v1","addr":"192.168.123.109:6807","nonce":2811766425}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":2811766425},{"type":"v1","addr":"192.168.123.109:6805","nonce":2811766425}]},"public_addr":"192.168.123.109:6801/2811766425","cluster_addr":"192.168.123.109:6803/2811766425","heartbeat_back_addr":"192.168.123.109:6807/2811766425","heartbeat_front_addr":"192.168.123.109:6805/2811766425","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:16:41.111365+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:17:14.101290+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:17:45.001792+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.100:6801/521837231":"2026-03-11T10:15:05.753536+0000","192.168.123.100:0/2092203":"2026-03-11T10:15:05.753536+0000","192.168.123.100:0/2293961577":"2026-03-11T10:15:05.753536+0000","192.168.123.100:0/1311426262":"2026-03-11T10:14:55.665189+0000","192.168.123.100:0/3326429106":"2026-03-11T10:14:55.665189+0000","192.168.123.100:0/2087321311":"2026-03-11T10:15:05.753536+0000","192.168.123.100:6800/521837231":"2026-03-11T10:15:05.753536+0000","192.168.123.100:0/1655622803":"2026-03-11T10:14:55.665189+0000","192.168.123.100:6801/3097082882":"2026-03-11T10:14:55.665189+0000","192.168.123.100:6800/3097082882":"2026-03-11T10:14:55.665189+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T10:18:15.554 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph tell osd.0 flush_pg_stats 2026-03-10T10:18:15.554 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph tell osd.1 flush_pg_stats 2026-03-10T10:18:15.554 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph tell osd.2 flush_pg_stats 2026-03-10T10:18:15.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:15 vm00 bash[20709]: cluster 2026-03-10T10:18:13.782133+0000 mgr.a (mgr.14150) 163 : cluster [DBG] pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:15.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:15 vm00 bash[20709]: cluster 2026-03-10T10:18:13.782133+0000 mgr.a (mgr.14150) 163 : cluster [DBG] pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:15.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:15 vm00 bash[20709]: audit 2026-03-10T10:18:15.494435+0000 mon.a (mon.0) 583 : audit [DBG] from='client.? 192.168.123.100:0/1399005341' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:18:15.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:15 vm00 bash[20709]: audit 2026-03-10T10:18:15.494435+0000 mon.a (mon.0) 583 : audit [DBG] from='client.? 192.168.123.100:0/1399005341' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:18:16.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:15 vm09 bash[20916]: cluster 2026-03-10T10:18:13.782133+0000 mgr.a (mgr.14150) 163 : cluster [DBG] pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:16.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:15 vm09 bash[20916]: cluster 2026-03-10T10:18:13.782133+0000 mgr.a (mgr.14150) 163 : cluster [DBG] pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:16.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:15 vm09 bash[20916]: audit 2026-03-10T10:18:15.494435+0000 mon.a (mon.0) 583 : audit [DBG] from='client.? 192.168.123.100:0/1399005341' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:18:16.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:15 vm09 bash[20916]: audit 2026-03-10T10:18:15.494435+0000 mon.a (mon.0) 583 : audit [DBG] from='client.? 192.168.123.100:0/1399005341' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:18:16.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:15 vm03 bash[21274]: cluster 2026-03-10T10:18:13.782133+0000 mgr.a (mgr.14150) 163 : cluster [DBG] pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:16.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:15 vm03 bash[21274]: cluster 2026-03-10T10:18:13.782133+0000 mgr.a (mgr.14150) 163 : cluster [DBG] pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:16.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:15 vm03 bash[21274]: audit 2026-03-10T10:18:15.494435+0000 mon.a (mon.0) 583 : audit [DBG] from='client.? 192.168.123.100:0/1399005341' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:18:16.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:15 vm03 bash[21274]: audit 2026-03-10T10:18:15.494435+0000 mon.a (mon.0) 583 : audit [DBG] from='client.? 192.168.123.100:0/1399005341' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:18:17.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:17 vm00 bash[20709]: cluster 2026-03-10T10:18:15.782442+0000 mgr.a (mgr.14150) 164 : cluster [DBG] pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:17.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:17 vm00 bash[20709]: cluster 2026-03-10T10:18:15.782442+0000 mgr.a (mgr.14150) 164 : cluster [DBG] pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:17.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:17 vm00 bash[20709]: audit 2026-03-10T10:18:17.404042+0000 mon.a (mon.0) 584 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:17.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:17 vm00 bash[20709]: audit 2026-03-10T10:18:17.404042+0000 mon.a (mon.0) 584 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:17.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:17 vm00 bash[20709]: audit 2026-03-10T10:18:17.409413+0000 mon.a (mon.0) 585 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:17.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:17 vm00 bash[20709]: audit 2026-03-10T10:18:17.409413+0000 mon.a (mon.0) 585 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:17.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:17 vm00 bash[20709]: audit 2026-03-10T10:18:17.414890+0000 mon.a (mon.0) 586 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:17.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:17 vm00 bash[20709]: audit 2026-03-10T10:18:17.414890+0000 mon.a (mon.0) 586 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:18.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:17 vm09 bash[20916]: cluster 2026-03-10T10:18:15.782442+0000 mgr.a (mgr.14150) 164 : cluster [DBG] pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:18.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:17 vm09 bash[20916]: cluster 2026-03-10T10:18:15.782442+0000 mgr.a (mgr.14150) 164 : cluster [DBG] pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:18.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:17 vm09 bash[20916]: audit 2026-03-10T10:18:17.404042+0000 mon.a (mon.0) 584 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:18.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:17 vm09 bash[20916]: audit 2026-03-10T10:18:17.404042+0000 mon.a (mon.0) 584 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:18.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:17 vm09 bash[20916]: audit 2026-03-10T10:18:17.409413+0000 mon.a (mon.0) 585 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:18.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:17 vm09 bash[20916]: audit 2026-03-10T10:18:17.409413+0000 mon.a (mon.0) 585 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:18.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:17 vm09 bash[20916]: audit 2026-03-10T10:18:17.414890+0000 mon.a (mon.0) 586 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:18.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:17 vm09 bash[20916]: audit 2026-03-10T10:18:17.414890+0000 mon.a (mon.0) 586 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:18.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:17 vm03 bash[21274]: cluster 2026-03-10T10:18:15.782442+0000 mgr.a (mgr.14150) 164 : cluster [DBG] pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:18.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:17 vm03 bash[21274]: cluster 2026-03-10T10:18:15.782442+0000 mgr.a (mgr.14150) 164 : cluster [DBG] pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:18.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:17 vm03 bash[21274]: audit 2026-03-10T10:18:17.404042+0000 mon.a (mon.0) 584 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:18.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:17 vm03 bash[21274]: audit 2026-03-10T10:18:17.404042+0000 mon.a (mon.0) 584 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:18.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:17 vm03 bash[21274]: audit 2026-03-10T10:18:17.409413+0000 mon.a (mon.0) 585 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:18.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:17 vm03 bash[21274]: audit 2026-03-10T10:18:17.409413+0000 mon.a (mon.0) 585 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:18.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:17 vm03 bash[21274]: audit 2026-03-10T10:18:17.414890+0000 mon.a (mon.0) 586 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:18.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:17 vm03 bash[21274]: audit 2026-03-10T10:18:17.414890+0000 mon.a (mon.0) 586 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.219 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:18:19.403 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:19 vm00 bash[20709]: cluster 2026-03-10T10:18:17.782722+0000 mgr.a (mgr.14150) 165 : cluster [DBG] pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:19.403 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:19 vm00 bash[20709]: cluster 2026-03-10T10:18:17.782722+0000 mgr.a (mgr.14150) 165 : cluster [DBG] pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:19.403 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:19 vm00 bash[20709]: audit 2026-03-10T10:18:17.946830+0000 mon.a (mon.0) 587 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.403 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:19 vm00 bash[20709]: audit 2026-03-10T10:18:17.946830+0000 mon.a (mon.0) 587 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.403 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:19 vm00 bash[20709]: audit 2026-03-10T10:18:17.953609+0000 mon.a (mon.0) 588 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.403 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:19 vm00 bash[20709]: audit 2026-03-10T10:18:17.953609+0000 mon.a (mon.0) 588 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.403 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:19 vm00 bash[20709]: audit 2026-03-10T10:18:17.960074+0000 mon.a (mon.0) 589 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.403 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:19 vm00 bash[20709]: audit 2026-03-10T10:18:17.960074+0000 mon.a (mon.0) 589 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:19 vm09 bash[20916]: cluster 2026-03-10T10:18:17.782722+0000 mgr.a (mgr.14150) 165 : cluster [DBG] pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:19.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:19 vm09 bash[20916]: cluster 2026-03-10T10:18:17.782722+0000 mgr.a (mgr.14150) 165 : cluster [DBG] pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:19.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:19 vm09 bash[20916]: audit 2026-03-10T10:18:17.946830+0000 mon.a (mon.0) 587 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:19 vm09 bash[20916]: audit 2026-03-10T10:18:17.946830+0000 mon.a (mon.0) 587 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:19 vm09 bash[20916]: audit 2026-03-10T10:18:17.953609+0000 mon.a (mon.0) 588 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:19 vm09 bash[20916]: audit 2026-03-10T10:18:17.953609+0000 mon.a (mon.0) 588 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:19 vm09 bash[20916]: audit 2026-03-10T10:18:17.960074+0000 mon.a (mon.0) 589 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:19 vm09 bash[20916]: audit 2026-03-10T10:18:17.960074+0000 mon.a (mon.0) 589 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.570 INFO:teuthology.orchestra.run.vm00.stdout:55834574863 2026-03-10T10:18:19.570 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph osd last-stat-seq osd.1 2026-03-10T10:18:19.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:19 vm03 bash[21274]: cluster 2026-03-10T10:18:17.782722+0000 mgr.a (mgr.14150) 165 : cluster [DBG] pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:19.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:19 vm03 bash[21274]: cluster 2026-03-10T10:18:17.782722+0000 mgr.a (mgr.14150) 165 : cluster [DBG] pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:19.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:19 vm03 bash[21274]: audit 2026-03-10T10:18:17.946830+0000 mon.a (mon.0) 587 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:19 vm03 bash[21274]: audit 2026-03-10T10:18:17.946830+0000 mon.a (mon.0) 587 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:19 vm03 bash[21274]: audit 2026-03-10T10:18:17.953609+0000 mon.a (mon.0) 588 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:19 vm03 bash[21274]: audit 2026-03-10T10:18:17.953609+0000 mon.a (mon.0) 588 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:19 vm03 bash[21274]: audit 2026-03-10T10:18:17.960074+0000 mon.a (mon.0) 589 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:19.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:19 vm03 bash[21274]: audit 2026-03-10T10:18:17.960074+0000 mon.a (mon.0) 589 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:20.224 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:18:20.225 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:18:20.546 INFO:teuthology.orchestra.run.vm00.stdout:77309411336 2026-03-10T10:18:20.547 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph osd last-stat-seq osd.2 2026-03-10T10:18:20.560 INFO:teuthology.orchestra.run.vm00.stdout:34359738389 2026-03-10T10:18:20.560 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph osd last-stat-seq osd.0 2026-03-10T10:18:21.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:21 vm00 bash[20709]: cluster 2026-03-10T10:18:19.783025+0000 mgr.a (mgr.14150) 166 : cluster [DBG] pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:21.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:21 vm00 bash[20709]: cluster 2026-03-10T10:18:19.783025+0000 mgr.a (mgr.14150) 166 : cluster [DBG] pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:21.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:21 vm09 bash[20916]: cluster 2026-03-10T10:18:19.783025+0000 mgr.a (mgr.14150) 166 : cluster [DBG] pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:21.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:21 vm09 bash[20916]: cluster 2026-03-10T10:18:19.783025+0000 mgr.a (mgr.14150) 166 : cluster [DBG] pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:21.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:21 vm03 bash[21274]: cluster 2026-03-10T10:18:19.783025+0000 mgr.a (mgr.14150) 166 : cluster [DBG] pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:21.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:21 vm03 bash[21274]: cluster 2026-03-10T10:18:19.783025+0000 mgr.a (mgr.14150) 166 : cluster [DBG] pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:23.233 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:18:23.579 INFO:teuthology.orchestra.run.vm00.stdout:55834574864 2026-03-10T10:18:23.643 INFO:tasks.cephadm.ceph_manager.ceph:need seq 55834574863 got 55834574864 for osd.1 2026-03-10T10:18:23.643 DEBUG:teuthology.parallel:result is None 2026-03-10T10:18:23.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:23 vm03 bash[21274]: cluster 2026-03-10T10:18:21.783360+0000 mgr.a (mgr.14150) 167 : cluster [DBG] pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:23.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:23 vm03 bash[21274]: cluster 2026-03-10T10:18:21.783360+0000 mgr.a (mgr.14150) 167 : cluster [DBG] pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:23.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:23 vm03 bash[21274]: audit 2026-03-10T10:18:22.498333+0000 mon.a (mon.0) 590 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:23.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:23 vm03 bash[21274]: audit 2026-03-10T10:18:22.498333+0000 mon.a (mon.0) 590 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:23.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:23 vm03 bash[21274]: audit 2026-03-10T10:18:22.503059+0000 mon.a (mon.0) 591 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:23.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:23 vm03 bash[21274]: audit 2026-03-10T10:18:22.503059+0000 mon.a (mon.0) 591 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:23.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:23 vm03 bash[21274]: audit 2026-03-10T10:18:22.507726+0000 mon.a (mon.0) 592 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:23.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:23 vm03 bash[21274]: audit 2026-03-10T10:18:22.507726+0000 mon.a (mon.0) 592 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:23.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:23 vm00 bash[20709]: cluster 2026-03-10T10:18:21.783360+0000 mgr.a (mgr.14150) 167 : cluster [DBG] pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:23.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:23 vm00 bash[20709]: cluster 2026-03-10T10:18:21.783360+0000 mgr.a (mgr.14150) 167 : cluster [DBG] pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:23.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:23 vm00 bash[20709]: audit 2026-03-10T10:18:22.498333+0000 mon.a (mon.0) 590 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:23.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:23 vm00 bash[20709]: audit 2026-03-10T10:18:22.498333+0000 mon.a (mon.0) 590 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:23.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:23 vm00 bash[20709]: audit 2026-03-10T10:18:22.503059+0000 mon.a (mon.0) 591 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:23.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:23 vm00 bash[20709]: audit 2026-03-10T10:18:22.503059+0000 mon.a (mon.0) 591 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:23.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:23 vm00 bash[20709]: audit 2026-03-10T10:18:22.507726+0000 mon.a (mon.0) 592 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:23.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:23 vm00 bash[20709]: audit 2026-03-10T10:18:22.507726+0000 mon.a (mon.0) 592 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:24.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:23 vm09 bash[20916]: cluster 2026-03-10T10:18:21.783360+0000 mgr.a (mgr.14150) 167 : cluster [DBG] pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:24.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:23 vm09 bash[20916]: cluster 2026-03-10T10:18:21.783360+0000 mgr.a (mgr.14150) 167 : cluster [DBG] pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:24.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:23 vm09 bash[20916]: audit 2026-03-10T10:18:22.498333+0000 mon.a (mon.0) 590 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:24.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:23 vm09 bash[20916]: audit 2026-03-10T10:18:22.498333+0000 mon.a (mon.0) 590 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:24.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:23 vm09 bash[20916]: audit 2026-03-10T10:18:22.503059+0000 mon.a (mon.0) 591 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:24.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:23 vm09 bash[20916]: audit 2026-03-10T10:18:22.503059+0000 mon.a (mon.0) 591 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:24.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:23 vm09 bash[20916]: audit 2026-03-10T10:18:22.507726+0000 mon.a (mon.0) 592 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:24.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:23 vm09 bash[20916]: audit 2026-03-10T10:18:22.507726+0000 mon.a (mon.0) 592 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:24.237 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:18:24.239 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:18:24.527 INFO:teuthology.orchestra.run.vm00.stdout:34359738389 2026-03-10T10:18:24.552 INFO:teuthology.orchestra.run.vm00.stdout:77309411337 2026-03-10T10:18:24.608 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738389 got 34359738389 for osd.0 2026-03-10T10:18:24.608 DEBUG:teuthology.parallel:result is None 2026-03-10T10:18:24.651 INFO:tasks.cephadm.ceph_manager.ceph:need seq 77309411336 got 77309411337 for osd.2 2026-03-10T10:18:24.651 DEBUG:teuthology.parallel:result is None 2026-03-10T10:18:24.651 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-10T10:18:24.651 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph pg dump --format=json 2026-03-10T10:18:24.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:24 vm03 bash[21274]: audit 2026-03-10T10:18:23.577678+0000 mon.c (mon.1) 14 : audit [DBG] from='client.? 192.168.123.100:0/493286387' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T10:18:24.854 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:24 vm03 bash[21274]: audit 2026-03-10T10:18:23.577678+0000 mon.c (mon.1) 14 : audit [DBG] from='client.? 192.168.123.100:0/493286387' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T10:18:24.854 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:24 vm03 bash[21274]: audit 2026-03-10T10:18:24.527172+0000 mon.b (mon.2) 8 : audit [DBG] from='client.? 192.168.123.100:0/4207223623' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T10:18:24.854 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:24 vm03 bash[21274]: audit 2026-03-10T10:18:24.527172+0000 mon.b (mon.2) 8 : audit [DBG] from='client.? 192.168.123.100:0/4207223623' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T10:18:24.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:24 vm00 bash[20709]: audit 2026-03-10T10:18:23.577678+0000 mon.c (mon.1) 14 : audit [DBG] from='client.? 192.168.123.100:0/493286387' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T10:18:24.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:24 vm00 bash[20709]: audit 2026-03-10T10:18:23.577678+0000 mon.c (mon.1) 14 : audit [DBG] from='client.? 192.168.123.100:0/493286387' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T10:18:24.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:24 vm00 bash[20709]: audit 2026-03-10T10:18:24.527172+0000 mon.b (mon.2) 8 : audit [DBG] from='client.? 192.168.123.100:0/4207223623' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T10:18:24.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:24 vm00 bash[20709]: audit 2026-03-10T10:18:24.527172+0000 mon.b (mon.2) 8 : audit [DBG] from='client.? 192.168.123.100:0/4207223623' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T10:18:25.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:24 vm09 bash[20916]: audit 2026-03-10T10:18:23.577678+0000 mon.c (mon.1) 14 : audit [DBG] from='client.? 192.168.123.100:0/493286387' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T10:18:25.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:24 vm09 bash[20916]: audit 2026-03-10T10:18:23.577678+0000 mon.c (mon.1) 14 : audit [DBG] from='client.? 192.168.123.100:0/493286387' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T10:18:25.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:24 vm09 bash[20916]: audit 2026-03-10T10:18:24.527172+0000 mon.b (mon.2) 8 : audit [DBG] from='client.? 192.168.123.100:0/4207223623' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T10:18:25.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:24 vm09 bash[20916]: audit 2026-03-10T10:18:24.527172+0000 mon.b (mon.2) 8 : audit [DBG] from='client.? 192.168.123.100:0/4207223623' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T10:18:25.819 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:25 vm09 bash[20916]: cluster 2026-03-10T10:18:23.783740+0000 mgr.a (mgr.14150) 168 : cluster [DBG] pgmap v111: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:25.819 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:25 vm09 bash[20916]: cluster 2026-03-10T10:18:23.783740+0000 mgr.a (mgr.14150) 168 : cluster [DBG] pgmap v111: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:25.819 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:25 vm09 bash[20916]: audit 2026-03-10T10:18:24.551752+0000 mon.b (mon.2) 9 : audit [DBG] from='client.? 192.168.123.100:0/2015893419' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T10:18:25.820 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:25 vm09 bash[20916]: audit 2026-03-10T10:18:24.551752+0000 mon.b (mon.2) 9 : audit [DBG] from='client.? 192.168.123.100:0/2015893419' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T10:18:25.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:25 vm03 bash[21274]: cluster 2026-03-10T10:18:23.783740+0000 mgr.a (mgr.14150) 168 : cluster [DBG] pgmap v111: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:25.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:25 vm03 bash[21274]: cluster 2026-03-10T10:18:23.783740+0000 mgr.a (mgr.14150) 168 : cluster [DBG] pgmap v111: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:25.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:25 vm03 bash[21274]: audit 2026-03-10T10:18:24.551752+0000 mon.b (mon.2) 9 : audit [DBG] from='client.? 192.168.123.100:0/2015893419' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T10:18:25.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:25 vm03 bash[21274]: audit 2026-03-10T10:18:24.551752+0000 mon.b (mon.2) 9 : audit [DBG] from='client.? 192.168.123.100:0/2015893419' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T10:18:25.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:25 vm00 bash[20709]: cluster 2026-03-10T10:18:23.783740+0000 mgr.a (mgr.14150) 168 : cluster [DBG] pgmap v111: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:25.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:25 vm00 bash[20709]: cluster 2026-03-10T10:18:23.783740+0000 mgr.a (mgr.14150) 168 : cluster [DBG] pgmap v111: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:25.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:25 vm00 bash[20709]: audit 2026-03-10T10:18:24.551752+0000 mon.b (mon.2) 9 : audit [DBG] from='client.? 192.168.123.100:0/2015893419' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T10:18:25.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:25 vm00 bash[20709]: audit 2026-03-10T10:18:24.551752+0000 mon.b (mon.2) 9 : audit [DBG] from='client.? 192.168.123.100:0/2015893419' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T10:18:27.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:27 vm03 bash[21274]: cluster 2026-03-10T10:18:25.784077+0000 mgr.a (mgr.14150) 169 : cluster [DBG] pgmap v112: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:27.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:27 vm03 bash[21274]: cluster 2026-03-10T10:18:25.784077+0000 mgr.a (mgr.14150) 169 : cluster [DBG] pgmap v112: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:27.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:27 vm00 bash[20709]: cluster 2026-03-10T10:18:25.784077+0000 mgr.a (mgr.14150) 169 : cluster [DBG] pgmap v112: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:27.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:27 vm00 bash[20709]: cluster 2026-03-10T10:18:25.784077+0000 mgr.a (mgr.14150) 169 : cluster [DBG] pgmap v112: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:28.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:27 vm09 bash[20916]: cluster 2026-03-10T10:18:25.784077+0000 mgr.a (mgr.14150) 169 : cluster [DBG] pgmap v112: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:28.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:27 vm09 bash[20916]: cluster 2026-03-10T10:18:25.784077+0000 mgr.a (mgr.14150) 169 : cluster [DBG] pgmap v112: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:29.273 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:18:29.551 INFO:teuthology.orchestra.run.vm00.stderr:dumped all 2026-03-10T10:18:29.551 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:18:29.620 INFO:teuthology.orchestra.run.vm00.stdout:{"pg_ready":true,"pg_map":{"version":113,"stamp":"2026-03-10T10:18:27.784251+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":3,"num_per_pool_osds":3,"num_per_pool_omap_osds":3,"kb":62902272,"kb_used":82812,"kb_used_data":1908,"kb_used_omap":4,"kb_used_meta":80443,"kb_avail":62819460,"statfs":{"total":64411926528,"available":64327127040,"internally_reserved":0,"allocated":1953792,"data_stored":1550235,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":4770,"internal_metadata":82373982},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"12.001936"},"pg_stats":[{"pgid":"1.0","version":"21'32","reported_seq":57,"reported_epoch":22,"state":"active+clean","last_fresh":"2026-03-10T10:17:51.019935+0000","last_change":"2026-03-10T10:17:50.130936+0000","last_active":"2026-03-10T10:17:51.019935+0000","last_peered":"2026-03-10T10:17:51.019935+0000","last_clean":"2026-03-10T10:17:51.019935+0000","last_became_active":"2026-03-10T10:17:50.130616+0000","last_became_peered":"2026-03-10T10:17:50.130616+0000","last_unstale":"2026-03-10T10:17:51.019935+0000","last_undegraded":"2026-03-10T10:17:51.019935+0000","last_fullsized":"2026-03-10T10:17:51.019935+0000","mapping_epoch":20,"log_start":"0'0","ondisk_log_start":"0'0","created":20,"last_epoch_clean":21,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T10:17:48.767031+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T10:17:48.767031+0000","last_clean_scrub_stamp":"2026-03-10T10:17:48.767031+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:14:09.713737+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,0],"acting":[1,2,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":2,"up_from":18,"seq":77309411338,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27604,"kb_used_data":636,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939820,"statfs":{"total":21470642176,"available":21442375680,"internally_reserved":0,"allocated":651264,"data_stored":516745,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":13,"seq":55834574865,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27604,"kb_used_data":636,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939820,"statfs":{"total":21470642176,"available":21442375680,"internally_reserved":0,"allocated":651264,"data_stored":516745,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738390,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27604,"kb_used_data":636,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939820,"statfs":{"total":21470642176,"available":21442375680,"internally_reserved":0,"allocated":651264,"data_stored":516745,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-10T10:18:29.620 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph pg dump --format=json 2026-03-10T10:18:29.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:29 vm03 bash[21274]: cluster 2026-03-10T10:18:27.784394+0000 mgr.a (mgr.14150) 170 : cluster [DBG] pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:29.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:29 vm03 bash[21274]: cluster 2026-03-10T10:18:27.784394+0000 mgr.a (mgr.14150) 170 : cluster [DBG] pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:29.886 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:29 vm00 bash[20709]: cluster 2026-03-10T10:18:27.784394+0000 mgr.a (mgr.14150) 170 : cluster [DBG] pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:29.886 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:29 vm00 bash[20709]: cluster 2026-03-10T10:18:27.784394+0000 mgr.a (mgr.14150) 170 : cluster [DBG] pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:30.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:29 vm09 bash[20916]: cluster 2026-03-10T10:18:27.784394+0000 mgr.a (mgr.14150) 170 : cluster [DBG] pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:30.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:29 vm09 bash[20916]: cluster 2026-03-10T10:18:27.784394+0000 mgr.a (mgr.14150) 170 : cluster [DBG] pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:31.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:30 vm00 bash[20709]: audit 2026-03-10T10:18:29.551514+0000 mgr.a (mgr.14150) 171 : audit [DBG] from='client.24236 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:31.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:30 vm00 bash[20709]: audit 2026-03-10T10:18:29.551514+0000 mgr.a (mgr.14150) 171 : audit [DBG] from='client.24236 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:31.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:30 vm09 bash[20916]: audit 2026-03-10T10:18:29.551514+0000 mgr.a (mgr.14150) 171 : audit [DBG] from='client.24236 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:31.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:30 vm09 bash[20916]: audit 2026-03-10T10:18:29.551514+0000 mgr.a (mgr.14150) 171 : audit [DBG] from='client.24236 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:31.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:30 vm03 bash[21274]: audit 2026-03-10T10:18:29.551514+0000 mgr.a (mgr.14150) 171 : audit [DBG] from='client.24236 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:31.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:30 vm03 bash[21274]: audit 2026-03-10T10:18:29.551514+0000 mgr.a (mgr.14150) 171 : audit [DBG] from='client.24236 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:32.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:31 vm09 bash[20916]: cluster 2026-03-10T10:18:29.784637+0000 mgr.a (mgr.14150) 172 : cluster [DBG] pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:32.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:31 vm09 bash[20916]: cluster 2026-03-10T10:18:29.784637+0000 mgr.a (mgr.14150) 172 : cluster [DBG] pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:32.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:31 vm03 bash[21274]: cluster 2026-03-10T10:18:29.784637+0000 mgr.a (mgr.14150) 172 : cluster [DBG] pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:32.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:31 vm03 bash[21274]: cluster 2026-03-10T10:18:29.784637+0000 mgr.a (mgr.14150) 172 : cluster [DBG] pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:32.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:31 vm00 bash[20709]: cluster 2026-03-10T10:18:29.784637+0000 mgr.a (mgr.14150) 172 : cluster [DBG] pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:32.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:31 vm00 bash[20709]: cluster 2026-03-10T10:18:29.784637+0000 mgr.a (mgr.14150) 172 : cluster [DBG] pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:33.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:32 vm09 bash[20916]: cluster 2026-03-10T10:18:31.784877+0000 mgr.a (mgr.14150) 173 : cluster [DBG] pgmap v115: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:33.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:32 vm09 bash[20916]: cluster 2026-03-10T10:18:31.784877+0000 mgr.a (mgr.14150) 173 : cluster [DBG] pgmap v115: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:33.285 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:18:33.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:32 vm03 bash[21274]: cluster 2026-03-10T10:18:31.784877+0000 mgr.a (mgr.14150) 173 : cluster [DBG] pgmap v115: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:33.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:32 vm03 bash[21274]: cluster 2026-03-10T10:18:31.784877+0000 mgr.a (mgr.14150) 173 : cluster [DBG] pgmap v115: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:33.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:32 vm00 bash[20709]: cluster 2026-03-10T10:18:31.784877+0000 mgr.a (mgr.14150) 173 : cluster [DBG] pgmap v115: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:33.432 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:32 vm00 bash[20709]: cluster 2026-03-10T10:18:31.784877+0000 mgr.a (mgr.14150) 173 : cluster [DBG] pgmap v115: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:33.902 INFO:teuthology.orchestra.run.vm00.stderr:dumped all 2026-03-10T10:18:33.902 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:18:33.979 INFO:teuthology.orchestra.run.vm00.stdout:{"pg_ready":true,"pg_map":{"version":116,"stamp":"2026-03-10T10:18:33.785034+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":3,"num_per_pool_osds":3,"num_per_pool_omap_osds":3,"kb":62902272,"kb_used":82812,"kb_used_data":1908,"kb_used_omap":4,"kb_used_meta":80443,"kb_avail":62819460,"statfs":{"total":64411926528,"available":64327127040,"internally_reserved":0,"allocated":1953792,"data_stored":1550235,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":4770,"internal_metadata":82373982},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"12.001846"},"pg_stats":[{"pgid":"1.0","version":"21'32","reported_seq":57,"reported_epoch":22,"state":"active+clean","last_fresh":"2026-03-10T10:17:51.019935+0000","last_change":"2026-03-10T10:17:50.130936+0000","last_active":"2026-03-10T10:17:51.019935+0000","last_peered":"2026-03-10T10:17:51.019935+0000","last_clean":"2026-03-10T10:17:51.019935+0000","last_became_active":"2026-03-10T10:17:50.130616+0000","last_became_peered":"2026-03-10T10:17:50.130616+0000","last_unstale":"2026-03-10T10:17:51.019935+0000","last_undegraded":"2026-03-10T10:17:51.019935+0000","last_fullsized":"2026-03-10T10:17:51.019935+0000","mapping_epoch":20,"log_start":"0'0","ondisk_log_start":"0'0","created":20,"last_epoch_clean":21,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T10:17:48.767031+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T10:17:48.767031+0000","last_clean_scrub_stamp":"2026-03-10T10:17:48.767031+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:14:09.713737+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,0],"acting":[1,2,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":2,"up_from":18,"seq":77309411339,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27604,"kb_used_data":636,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939820,"statfs":{"total":21470642176,"available":21442375680,"internally_reserved":0,"allocated":651264,"data_stored":516745,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":13,"seq":55834574866,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27604,"kb_used_data":636,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939820,"statfs":{"total":21470642176,"available":21442375680,"internally_reserved":0,"allocated":651264,"data_stored":516745,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738392,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27604,"kb_used_data":636,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939820,"statfs":{"total":21470642176,"available":21442375680,"internally_reserved":0,"allocated":651264,"data_stored":516745,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-10T10:18:33.979 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-10T10:18:33.979 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-10T10:18:33.979 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-10T10:18:33.980 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph health --format=json 2026-03-10T10:18:35.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:34 vm09 bash[20916]: cluster 2026-03-10T10:18:33.785171+0000 mgr.a (mgr.14150) 174 : cluster [DBG] pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:35.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:34 vm09 bash[20916]: cluster 2026-03-10T10:18:33.785171+0000 mgr.a (mgr.14150) 174 : cluster [DBG] pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:35.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:34 vm09 bash[20916]: audit 2026-03-10T10:18:33.902470+0000 mgr.a (mgr.14150) 175 : audit [DBG] from='client.14367 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:35.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:34 vm09 bash[20916]: audit 2026-03-10T10:18:33.902470+0000 mgr.a (mgr.14150) 175 : audit [DBG] from='client.14367 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:35.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:34 vm03 bash[21274]: cluster 2026-03-10T10:18:33.785171+0000 mgr.a (mgr.14150) 174 : cluster [DBG] pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:35.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:34 vm03 bash[21274]: cluster 2026-03-10T10:18:33.785171+0000 mgr.a (mgr.14150) 174 : cluster [DBG] pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:35.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:34 vm03 bash[21274]: audit 2026-03-10T10:18:33.902470+0000 mgr.a (mgr.14150) 175 : audit [DBG] from='client.14367 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:35.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:34 vm03 bash[21274]: audit 2026-03-10T10:18:33.902470+0000 mgr.a (mgr.14150) 175 : audit [DBG] from='client.14367 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:35.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:34 vm00 bash[20709]: cluster 2026-03-10T10:18:33.785171+0000 mgr.a (mgr.14150) 174 : cluster [DBG] pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:35.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:34 vm00 bash[20709]: cluster 2026-03-10T10:18:33.785171+0000 mgr.a (mgr.14150) 174 : cluster [DBG] pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:35.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:34 vm00 bash[20709]: audit 2026-03-10T10:18:33.902470+0000 mgr.a (mgr.14150) 175 : audit [DBG] from='client.14367 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:35.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:34 vm00 bash[20709]: audit 2026-03-10T10:18:33.902470+0000 mgr.a (mgr.14150) 175 : audit [DBG] from='client.14367 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:37.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:36 vm09 bash[20916]: cluster 2026-03-10T10:18:35.785416+0000 mgr.a (mgr.14150) 176 : cluster [DBG] pgmap v117: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:37.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:36 vm09 bash[20916]: cluster 2026-03-10T10:18:35.785416+0000 mgr.a (mgr.14150) 176 : cluster [DBG] pgmap v117: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:37.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:36 vm03 bash[21274]: cluster 2026-03-10T10:18:35.785416+0000 mgr.a (mgr.14150) 176 : cluster [DBG] pgmap v117: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:37.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:36 vm03 bash[21274]: cluster 2026-03-10T10:18:35.785416+0000 mgr.a (mgr.14150) 176 : cluster [DBG] pgmap v117: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:37.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:36 vm00 bash[20709]: cluster 2026-03-10T10:18:35.785416+0000 mgr.a (mgr.14150) 176 : cluster [DBG] pgmap v117: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:37.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:36 vm00 bash[20709]: cluster 2026-03-10T10:18:35.785416+0000 mgr.a (mgr.14150) 176 : cluster [DBG] pgmap v117: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:38.605 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:18:38.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:38 vm09 bash[20916]: audit 2026-03-10T10:18:37.482763+0000 mon.a (mon.0) 593 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:38 vm09 bash[20916]: audit 2026-03-10T10:18:37.482763+0000 mon.a (mon.0) 593 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:38 vm09 bash[20916]: audit 2026-03-10T10:18:37.488394+0000 mon.a (mon.0) 594 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:38 vm09 bash[20916]: audit 2026-03-10T10:18:37.488394+0000 mon.a (mon.0) 594 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:38 vm09 bash[20916]: audit 2026-03-10T10:18:37.495199+0000 mon.a (mon.0) 595 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:38 vm09 bash[20916]: audit 2026-03-10T10:18:37.495199+0000 mon.a (mon.0) 595 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:38 vm09 bash[20916]: audit 2026-03-10T10:18:38.025116+0000 mon.a (mon.0) 596 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:38 vm09 bash[20916]: audit 2026-03-10T10:18:38.025116+0000 mon.a (mon.0) 596 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:38 vm09 bash[20916]: audit 2026-03-10T10:18:38.030552+0000 mon.a (mon.0) 597 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:38 vm09 bash[20916]: audit 2026-03-10T10:18:38.030552+0000 mon.a (mon.0) 597 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:38 vm09 bash[20916]: audit 2026-03-10T10:18:38.036539+0000 mon.a (mon.0) 598 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:38 vm09 bash[20916]: audit 2026-03-10T10:18:38.036539+0000 mon.a (mon.0) 598 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:38 vm03 bash[21274]: audit 2026-03-10T10:18:37.482763+0000 mon.a (mon.0) 593 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:38 vm03 bash[21274]: audit 2026-03-10T10:18:37.482763+0000 mon.a (mon.0) 593 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:38 vm03 bash[21274]: audit 2026-03-10T10:18:37.488394+0000 mon.a (mon.0) 594 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:38 vm03 bash[21274]: audit 2026-03-10T10:18:37.488394+0000 mon.a (mon.0) 594 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:38 vm03 bash[21274]: audit 2026-03-10T10:18:37.495199+0000 mon.a (mon.0) 595 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:38 vm03 bash[21274]: audit 2026-03-10T10:18:37.495199+0000 mon.a (mon.0) 595 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:38 vm03 bash[21274]: audit 2026-03-10T10:18:38.025116+0000 mon.a (mon.0) 596 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:38 vm03 bash[21274]: audit 2026-03-10T10:18:38.025116+0000 mon.a (mon.0) 596 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:38 vm03 bash[21274]: audit 2026-03-10T10:18:38.030552+0000 mon.a (mon.0) 597 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:38 vm03 bash[21274]: audit 2026-03-10T10:18:38.030552+0000 mon.a (mon.0) 597 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:38 vm03 bash[21274]: audit 2026-03-10T10:18:38.036539+0000 mon.a (mon.0) 598 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:38 vm03 bash[21274]: audit 2026-03-10T10:18:38.036539+0000 mon.a (mon.0) 598 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.889 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:18:38.889 INFO:teuthology.orchestra.run.vm00.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-10T10:18:38.902 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:38 vm00 bash[20709]: audit 2026-03-10T10:18:37.482763+0000 mon.a (mon.0) 593 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.902 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:38 vm00 bash[20709]: audit 2026-03-10T10:18:37.482763+0000 mon.a (mon.0) 593 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.902 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:38 vm00 bash[20709]: audit 2026-03-10T10:18:37.488394+0000 mon.a (mon.0) 594 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.902 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:38 vm00 bash[20709]: audit 2026-03-10T10:18:37.488394+0000 mon.a (mon.0) 594 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.902 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:38 vm00 bash[20709]: audit 2026-03-10T10:18:37.495199+0000 mon.a (mon.0) 595 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.902 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:38 vm00 bash[20709]: audit 2026-03-10T10:18:37.495199+0000 mon.a (mon.0) 595 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.902 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:38 vm00 bash[20709]: audit 2026-03-10T10:18:38.025116+0000 mon.a (mon.0) 596 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.902 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:38 vm00 bash[20709]: audit 2026-03-10T10:18:38.025116+0000 mon.a (mon.0) 596 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.902 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:38 vm00 bash[20709]: audit 2026-03-10T10:18:38.030552+0000 mon.a (mon.0) 597 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.902 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:38 vm00 bash[20709]: audit 2026-03-10T10:18:38.030552+0000 mon.a (mon.0) 597 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.902 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:38 vm00 bash[20709]: audit 2026-03-10T10:18:38.036539+0000 mon.a (mon.0) 598 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.902 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:38 vm00 bash[20709]: audit 2026-03-10T10:18:38.036539+0000 mon.a (mon.0) 598 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:38.960 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-10T10:18:38.960 INFO:tasks.cephadm:Setup complete, yielding 2026-03-10T10:18:38.960 INFO:teuthology.run_tasks:Running task cephadm.apply... 2026-03-10T10:18:38.963 INFO:tasks.cephadm:Applying spec(s): placement: count: 3 service_id: foo service_type: mon spec: crush_locations: host.a: - datacenter=a host.b: - datacenter=b - rack=2 host.c: - datacenter=a - rack=3 2026-03-10T10:18:38.963 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph orch apply -i - 2026-03-10T10:18:39.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:39 vm09 bash[20916]: cluster 2026-03-10T10:18:37.785703+0000 mgr.a (mgr.14150) 177 : cluster [DBG] pgmap v118: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:39.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:39 vm09 bash[20916]: cluster 2026-03-10T10:18:37.785703+0000 mgr.a (mgr.14150) 177 : cluster [DBG] pgmap v118: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:39.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:39 vm09 bash[20916]: audit 2026-03-10T10:18:38.890271+0000 mon.a (mon.0) 599 : audit [DBG] from='client.? 192.168.123.100:0/3001636994' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T10:18:39.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:39 vm09 bash[20916]: audit 2026-03-10T10:18:38.890271+0000 mon.a (mon.0) 599 : audit [DBG] from='client.? 192.168.123.100:0/3001636994' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T10:18:39.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:39 vm03 bash[21274]: cluster 2026-03-10T10:18:37.785703+0000 mgr.a (mgr.14150) 177 : cluster [DBG] pgmap v118: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:39.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:39 vm03 bash[21274]: cluster 2026-03-10T10:18:37.785703+0000 mgr.a (mgr.14150) 177 : cluster [DBG] pgmap v118: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:39.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:39 vm03 bash[21274]: audit 2026-03-10T10:18:38.890271+0000 mon.a (mon.0) 599 : audit [DBG] from='client.? 192.168.123.100:0/3001636994' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T10:18:39.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:39 vm03 bash[21274]: audit 2026-03-10T10:18:38.890271+0000 mon.a (mon.0) 599 : audit [DBG] from='client.? 192.168.123.100:0/3001636994' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T10:18:39.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:39 vm00 bash[20709]: cluster 2026-03-10T10:18:37.785703+0000 mgr.a (mgr.14150) 177 : cluster [DBG] pgmap v118: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:39.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:39 vm00 bash[20709]: cluster 2026-03-10T10:18:37.785703+0000 mgr.a (mgr.14150) 177 : cluster [DBG] pgmap v118: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:39.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:39 vm00 bash[20709]: audit 2026-03-10T10:18:38.890271+0000 mon.a (mon.0) 599 : audit [DBG] from='client.? 192.168.123.100:0/3001636994' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T10:18:39.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:39 vm00 bash[20709]: audit 2026-03-10T10:18:38.890271+0000 mon.a (mon.0) 599 : audit [DBG] from='client.? 192.168.123.100:0/3001636994' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T10:18:41.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:41 vm09 bash[20916]: cluster 2026-03-10T10:18:39.786041+0000 mgr.a (mgr.14150) 178 : cluster [DBG] pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:41.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:41 vm09 bash[20916]: cluster 2026-03-10T10:18:39.786041+0000 mgr.a (mgr.14150) 178 : cluster [DBG] pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:41.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:41 vm03 bash[21274]: cluster 2026-03-10T10:18:39.786041+0000 mgr.a (mgr.14150) 178 : cluster [DBG] pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:41.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:41 vm03 bash[21274]: cluster 2026-03-10T10:18:39.786041+0000 mgr.a (mgr.14150) 178 : cluster [DBG] pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:41.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:41 vm00 bash[20709]: cluster 2026-03-10T10:18:39.786041+0000 mgr.a (mgr.14150) 178 : cluster [DBG] pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:41.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:41 vm00 bash[20709]: cluster 2026-03-10T10:18:39.786041+0000 mgr.a (mgr.14150) 178 : cluster [DBG] pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:42.620 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:18:42.913 INFO:teuthology.orchestra.run.vm00.stdout:Scheduled mon update... 2026-03-10T10:18:42.989 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T10:18:42.991 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm00.local 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- bash -c 'set -ex 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> # since we don'"'"'t know the real hostnames before the test, the next 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> # bit is in order to replace the fake hostnames "host.a/b/c" with 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> # the actual names cephadm knows the host by within the mon spec 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> ceph orch host ls --format json | jq -r '"'"'.[] | .hostname'"'"' > realnames 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> echo $'"'"'host.a\nhost.b\nhost.c'"'"' > fakenames 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> echo $'"'"'a\nb\nc'"'"' > mon_ids 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> echo $'"'"'{datacenter=a}\n{datacenter=b,rack=2}\n{datacenter=a,rack=3}'"'"' > crush_locs 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> ceph orch ls --service-name mon --export > mon.yaml 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> MONSPEC=`cat mon.yaml` 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> echo "$MONSPEC" 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> while read realname <&3 && read fakename <&4; do 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> MONSPEC="${MONSPEC//$fakename/$realname}" 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> done 3 echo "$MONSPEC" > mon.yaml 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> cat mon.yaml 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> # now the spec should have the real hostnames, so let'"'"'s re-apply 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> ceph orch apply -i mon.yaml 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> sleep 90 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> ceph orch ps --refresh 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> ceph orch ls --service-name mon --export > mon.yaml; ceph orch apply -i mon.yaml 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> sleep 90 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> ceph mon dump 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> ceph mon dump --format json 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> # verify all the crush locations got set from "ceph mon dump" output 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> while read monid <&3 && read crushloc <&4; do 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> ceph mon dump --format json | jq --arg monid "$monid" --arg crushloc "$crushloc" -e '"'"'.mons | .[] | select(.name == $monid) | .crush_location == $crushloc'"'"' 2026-03-10T10:18:42.991 DEBUG:teuthology.orchestra.run.vm00:> done 3 ' 2026-03-10T10:18:43.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: cluster 2026-03-10T10:18:41.786312+0000 mgr.a (mgr.14150) 179 : cluster [DBG] pgmap v120: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:43.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: cluster 2026-03-10T10:18:41.786312+0000 mgr.a (mgr.14150) 179 : cluster [DBG] pgmap v120: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:43.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.566441+0000 mon.a (mon.0) 600 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:43.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.566441+0000 mon.a (mon.0) 600 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:43.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.569879+0000 mon.a (mon.0) 601 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:43.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.569879+0000 mon.a (mon.0) 601 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:43.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.573652+0000 mon.a (mon.0) 602 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:43.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.573652+0000 mon.a (mon.0) 602 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:43.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.913066+0000 mon.a (mon.0) 603 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:43.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.913066+0000 mon.a (mon.0) 603 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:43.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.913835+0000 mon.a (mon.0) 604 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:43.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.913835+0000 mon.a (mon.0) 604 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:43.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.915700+0000 mon.a (mon.0) 605 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:43.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.915700+0000 mon.a (mon.0) 605 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:43.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.916252+0000 mon.a (mon.0) 606 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:43.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.916252+0000 mon.a (mon.0) 606 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:43.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.921011+0000 mon.a (mon.0) 607 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:43.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.921011+0000 mon.a (mon.0) 607 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:43.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.922402+0000 mon.a (mon.0) 608 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-10T10:18:43.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.922402+0000 mon.a (mon.0) 608 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-10T10:18:43.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.926101+0000 mon.a (mon.0) 609 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:43.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:43 vm00 bash[20709]: audit 2026-03-10T10:18:42.926101+0000 mon.a (mon.0) 609 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: cluster 2026-03-10T10:18:41.786312+0000 mgr.a (mgr.14150) 179 : cluster [DBG] pgmap v120: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: cluster 2026-03-10T10:18:41.786312+0000 mgr.a (mgr.14150) 179 : cluster [DBG] pgmap v120: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.566441+0000 mon.a (mon.0) 600 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.566441+0000 mon.a (mon.0) 600 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.569879+0000 mon.a (mon.0) 601 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.569879+0000 mon.a (mon.0) 601 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.573652+0000 mon.a (mon.0) 602 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.573652+0000 mon.a (mon.0) 602 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.913066+0000 mon.a (mon.0) 603 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.913066+0000 mon.a (mon.0) 603 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.913835+0000 mon.a (mon.0) 604 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.913835+0000 mon.a (mon.0) 604 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.915700+0000 mon.a (mon.0) 605 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.915700+0000 mon.a (mon.0) 605 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.916252+0000 mon.a (mon.0) 606 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.916252+0000 mon.a (mon.0) 606 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.921011+0000 mon.a (mon.0) 607 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.921011+0000 mon.a (mon.0) 607 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.922402+0000 mon.a (mon.0) 608 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.922402+0000 mon.a (mon.0) 608 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.926101+0000 mon.a (mon.0) 609 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:43 vm09 bash[20916]: audit 2026-03-10T10:18:42.926101+0000 mon.a (mon.0) 609 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: cluster 2026-03-10T10:18:41.786312+0000 mgr.a (mgr.14150) 179 : cluster [DBG] pgmap v120: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: cluster 2026-03-10T10:18:41.786312+0000 mgr.a (mgr.14150) 179 : cluster [DBG] pgmap v120: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.566441+0000 mon.a (mon.0) 600 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.566441+0000 mon.a (mon.0) 600 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.569879+0000 mon.a (mon.0) 601 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.569879+0000 mon.a (mon.0) 601 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.573652+0000 mon.a (mon.0) 602 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.573652+0000 mon.a (mon.0) 602 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.913066+0000 mon.a (mon.0) 603 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.913066+0000 mon.a (mon.0) 603 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.913835+0000 mon.a (mon.0) 604 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.913835+0000 mon.a (mon.0) 604 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.915700+0000 mon.a (mon.0) 605 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.915700+0000 mon.a (mon.0) 605 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.916252+0000 mon.a (mon.0) 606 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.916252+0000 mon.a (mon.0) 606 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.921011+0000 mon.a (mon.0) 607 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.921011+0000 mon.a (mon.0) 607 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.922402+0000 mon.a (mon.0) 608 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.922402+0000 mon.a (mon.0) 608 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.926101+0000 mon.a (mon.0) 609 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:43 vm03 bash[21274]: audit 2026-03-10T10:18:42.926101+0000 mon.a (mon.0) 609 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:44.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:44 vm00 bash[20709]: audit 2026-03-10T10:18:42.906962+0000 mgr.a (mgr.14150) 180 : audit [DBG] from='client.14379 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:44.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:44 vm00 bash[20709]: audit 2026-03-10T10:18:42.906962+0000 mgr.a (mgr.14150) 180 : audit [DBG] from='client.14379 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:44.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:44 vm00 bash[20709]: cephadm 2026-03-10T10:18:42.908546+0000 mgr.a (mgr.14150) 181 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:18:44.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:44 vm00 bash[20709]: cephadm 2026-03-10T10:18:42.908546+0000 mgr.a (mgr.14150) 181 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:18:45.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:44 vm09 bash[20916]: audit 2026-03-10T10:18:42.906962+0000 mgr.a (mgr.14150) 180 : audit [DBG] from='client.14379 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:45.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:44 vm09 bash[20916]: audit 2026-03-10T10:18:42.906962+0000 mgr.a (mgr.14150) 180 : audit [DBG] from='client.14379 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:45.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:44 vm09 bash[20916]: cephadm 2026-03-10T10:18:42.908546+0000 mgr.a (mgr.14150) 181 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:18:45.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:44 vm09 bash[20916]: cephadm 2026-03-10T10:18:42.908546+0000 mgr.a (mgr.14150) 181 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:18:45.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:44 vm03 bash[21274]: audit 2026-03-10T10:18:42.906962+0000 mgr.a (mgr.14150) 180 : audit [DBG] from='client.14379 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:45.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:44 vm03 bash[21274]: audit 2026-03-10T10:18:42.906962+0000 mgr.a (mgr.14150) 180 : audit [DBG] from='client.14379 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:45.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:44 vm03 bash[21274]: cephadm 2026-03-10T10:18:42.908546+0000 mgr.a (mgr.14150) 181 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:18:45.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:44 vm03 bash[21274]: cephadm 2026-03-10T10:18:42.908546+0000 mgr.a (mgr.14150) 181 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:18:45.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:45 vm00 bash[20709]: cluster 2026-03-10T10:18:43.786611+0000 mgr.a (mgr.14150) 182 : cluster [DBG] pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:45.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:45 vm00 bash[20709]: cluster 2026-03-10T10:18:43.786611+0000 mgr.a (mgr.14150) 182 : cluster [DBG] pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:46.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:45 vm09 bash[20916]: cluster 2026-03-10T10:18:43.786611+0000 mgr.a (mgr.14150) 182 : cluster [DBG] pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:46.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:45 vm09 bash[20916]: cluster 2026-03-10T10:18:43.786611+0000 mgr.a (mgr.14150) 182 : cluster [DBG] pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:46.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:45 vm03 bash[21274]: cluster 2026-03-10T10:18:43.786611+0000 mgr.a (mgr.14150) 182 : cluster [DBG] pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:46.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:45 vm03 bash[21274]: cluster 2026-03-10T10:18:43.786611+0000 mgr.a (mgr.14150) 182 : cluster [DBG] pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:46.634 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:18:46.757 INFO:teuthology.orchestra.run.vm00.stderr:+ ceph orch host ls --format json 2026-03-10T10:18:46.757 INFO:teuthology.orchestra.run.vm00.stderr:+ jq -r '.[] | .hostname' 2026-03-10T10:18:46.933 INFO:teuthology.orchestra.run.vm00.stderr:+ echo 'host.a 2026-03-10T10:18:46.933 INFO:teuthology.orchestra.run.vm00.stderr:host.b 2026-03-10T10:18:46.933 INFO:teuthology.orchestra.run.vm00.stderr:host.c' 2026-03-10T10:18:46.933 INFO:teuthology.orchestra.run.vm00.stderr:+ echo 'a 2026-03-10T10:18:46.933 INFO:teuthology.orchestra.run.vm00.stderr:b 2026-03-10T10:18:46.933 INFO:teuthology.orchestra.run.vm00.stderr:c' 2026-03-10T10:18:46.933 INFO:teuthology.orchestra.run.vm00.stderr:+ echo '{datacenter=a} 2026-03-10T10:18:46.933 INFO:teuthology.orchestra.run.vm00.stderr:{datacenter=b,rack=2} 2026-03-10T10:18:46.933 INFO:teuthology.orchestra.run.vm00.stderr:{datacenter=a,rack=3}' 2026-03-10T10:18:46.933 INFO:teuthology.orchestra.run.vm00.stderr:+ ceph orch ls --service-name mon --export 2026-03-10T10:18:47.103 INFO:teuthology.orchestra.run.vm00.stderr:++ cat mon.yaml 2026-03-10T10:18:47.103 INFO:teuthology.orchestra.run.vm00.stderr:+ MONSPEC='service_type: mon 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr:service_name: mon 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr:placement: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: count: 3 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr:spec: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: crush_locations: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: host.a: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=a 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: host.b: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=b 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: - rack=2 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: host.c: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=a 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: - rack=3' 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr:+ echo 'service_type: mon 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr:service_name: mon 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr:placement: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: count: 3 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr:spec: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: crush_locations: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: host.a: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=a 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: host.b: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=b 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: - rack=2 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: host.c: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=a 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stderr: - rack=3' 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stdout:service_type: mon 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stdout:service_name: mon 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stdout:placement: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stdout: count: 3 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stdout:spec: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stdout: crush_locations: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stdout: host.a: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stdout: - datacenter=a 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stdout: host.b: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stdout: - datacenter=b 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stdout: - rack=2 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stdout: host.c: 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stdout: - datacenter=a 2026-03-10T10:18:47.104 INFO:teuthology.orchestra.run.vm00.stdout: - rack=3 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:+ read realname 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:+ read fakename 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:+ MONSPEC='service_type: mon 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:service_name: mon 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:placement: 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: count: 3 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:spec: 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: crush_locations: 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: vm00: 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=a 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: host.b: 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=b 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: - rack=2 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: host.c: 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=a 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: - rack=3' 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:+ read realname 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:+ read fakename 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:+ MONSPEC='service_type: mon 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:service_name: mon 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:placement: 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: count: 3 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:spec: 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: crush_locations: 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: vm00: 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=a 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: vm03: 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=b 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: - rack=2 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: host.c: 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=a 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: - rack=3' 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:+ read realname 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:+ read fakename 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:+ MONSPEC='service_type: mon 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:service_name: mon 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:placement: 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: count: 3 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr:spec: 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: crush_locations: 2026-03-10T10:18:47.105 INFO:teuthology.orchestra.run.vm00.stderr: vm00: 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=a 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr: vm03: 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=b 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr: - rack=2 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr: vm09: 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=a 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr: - rack=3' 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr:+ read realname 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr:+ echo 'service_type: mon 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr:service_name: mon 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr:placement: 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr: count: 3 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr:spec: 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr: crush_locations: 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr: vm00: 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=a 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr: vm03: 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=b 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr: - rack=2 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr: vm09: 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr: - datacenter=a 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr: - rack=3' 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stderr:+ cat mon.yaml 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stdout:service_type: mon 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stdout:service_name: mon 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stdout:placement: 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stdout: count: 3 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stdout:spec: 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stdout: crush_locations: 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stdout: vm00: 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stdout: - datacenter=a 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stdout: vm03: 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stdout: - datacenter=b 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stdout: - rack=2 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stdout: vm09: 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stdout: - datacenter=a 2026-03-10T10:18:47.106 INFO:teuthology.orchestra.run.vm00.stdout: - rack=3 2026-03-10T10:18:47.107 INFO:teuthology.orchestra.run.vm00.stderr:+ ceph orch apply -i mon.yaml 2026-03-10T10:18:47.478 INFO:teuthology.orchestra.run.vm00.stdout:Scheduled mon update... 2026-03-10T10:18:47.491 INFO:teuthology.orchestra.run.vm00.stderr:+ sleep 90 2026-03-10T10:18:47.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:47 vm09 bash[20916]: cluster 2026-03-10T10:18:45.786903+0000 mgr.a (mgr.14150) 183 : cluster [DBG] pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:47.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:47 vm09 bash[20916]: cluster 2026-03-10T10:18:45.786903+0000 mgr.a (mgr.14150) 183 : cluster [DBG] pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:47.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:47 vm09 bash[20916]: audit 2026-03-10T10:18:47.477909+0000 mon.a (mon.0) 610 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:47.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:47 vm09 bash[20916]: audit 2026-03-10T10:18:47.477909+0000 mon.a (mon.0) 610 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:47.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:47 vm09 bash[20916]: audit 2026-03-10T10:18:47.479165+0000 mon.a (mon.0) 611 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:47.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:47 vm09 bash[20916]: audit 2026-03-10T10:18:47.479165+0000 mon.a (mon.0) 611 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:47.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:47 vm09 bash[20916]: audit 2026-03-10T10:18:47.481016+0000 mon.a (mon.0) 612 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:47.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:47 vm09 bash[20916]: audit 2026-03-10T10:18:47.481016+0000 mon.a (mon.0) 612 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:47.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:47 vm09 bash[20916]: audit 2026-03-10T10:18:47.481446+0000 mon.a (mon.0) 613 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:47.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:47 vm09 bash[20916]: audit 2026-03-10T10:18:47.481446+0000 mon.a (mon.0) 613 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:48.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:47 vm03 bash[21274]: cluster 2026-03-10T10:18:45.786903+0000 mgr.a (mgr.14150) 183 : cluster [DBG] pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:48.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:47 vm03 bash[21274]: cluster 2026-03-10T10:18:45.786903+0000 mgr.a (mgr.14150) 183 : cluster [DBG] pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:48.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:47 vm03 bash[21274]: audit 2026-03-10T10:18:47.477909+0000 mon.a (mon.0) 610 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:48.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:47 vm03 bash[21274]: audit 2026-03-10T10:18:47.477909+0000 mon.a (mon.0) 610 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:48.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:47 vm03 bash[21274]: audit 2026-03-10T10:18:47.479165+0000 mon.a (mon.0) 611 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:48.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:47 vm03 bash[21274]: audit 2026-03-10T10:18:47.479165+0000 mon.a (mon.0) 611 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:48.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:47 vm03 bash[21274]: audit 2026-03-10T10:18:47.481016+0000 mon.a (mon.0) 612 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:48.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:47 vm03 bash[21274]: audit 2026-03-10T10:18:47.481016+0000 mon.a (mon.0) 612 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:48.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:47 vm03 bash[21274]: audit 2026-03-10T10:18:47.481446+0000 mon.a (mon.0) 613 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:48.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:47 vm03 bash[21274]: audit 2026-03-10T10:18:47.481446+0000 mon.a (mon.0) 613 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:48.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:47 vm00 bash[20709]: cluster 2026-03-10T10:18:45.786903+0000 mgr.a (mgr.14150) 183 : cluster [DBG] pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:48.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:47 vm00 bash[20709]: cluster 2026-03-10T10:18:45.786903+0000 mgr.a (mgr.14150) 183 : cluster [DBG] pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:48.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:47 vm00 bash[20709]: audit 2026-03-10T10:18:47.477909+0000 mon.a (mon.0) 610 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:48.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:47 vm00 bash[20709]: audit 2026-03-10T10:18:47.477909+0000 mon.a (mon.0) 610 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:18:48.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:47 vm00 bash[20709]: audit 2026-03-10T10:18:47.479165+0000 mon.a (mon.0) 611 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:48.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:47 vm00 bash[20709]: audit 2026-03-10T10:18:47.479165+0000 mon.a (mon.0) 611 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:18:48.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:47 vm00 bash[20709]: audit 2026-03-10T10:18:47.481016+0000 mon.a (mon.0) 612 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:48.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:47 vm00 bash[20709]: audit 2026-03-10T10:18:47.481016+0000 mon.a (mon.0) 612 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:18:48.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:47 vm00 bash[20709]: audit 2026-03-10T10:18:47.481446+0000 mon.a (mon.0) 613 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:48.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:47 vm00 bash[20709]: audit 2026-03-10T10:18:47.481446+0000 mon.a (mon.0) 613 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:18:53.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: audit 2026-03-10T10:18:46.922198+0000 mgr.a (mgr.14150) 184 : audit [DBG] from='client.14385 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: audit 2026-03-10T10:18:46.922198+0000 mgr.a (mgr.14150) 184 : audit [DBG] from='client.14385 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: audit 2026-03-10T10:18:47.090343+0000 mgr.a (mgr.14150) 185 : audit [DBG] from='client.24268 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: audit 2026-03-10T10:18:47.090343+0000 mgr.a (mgr.14150) 185 : audit [DBG] from='client.24268 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: audit 2026-03-10T10:18:47.263971+0000 mgr.a (mgr.14150) 186 : audit [DBG] from='client.14397 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: audit 2026-03-10T10:18:47.263971+0000 mgr.a (mgr.14150) 186 : audit [DBG] from='client.14397 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cephadm 2026-03-10T10:18:47.265525+0000 mgr.a (mgr.14150) 187 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cephadm 2026-03-10T10:18:47.265525+0000 mgr.a (mgr.14150) 187 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cephadm 2026-03-10T10:18:47.675866+0000 mgr.a (mgr.14150) 188 : cephadm [INF] Setting crush location for mon a to {datacenter=a} 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cephadm 2026-03-10T10:18:47.675866+0000 mgr.a (mgr.14150) 188 : cephadm [INF] Setting crush location for mon a to {datacenter=a} 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: audit 2026-03-10T10:18:47.726201+0000 mon.a (mon.0) 617 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "a", "args": ["datacenter=a"]}]': finished 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: audit 2026-03-10T10:18:47.726201+0000 mon.a (mon.0) 617 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "a", "args": ["datacenter=a"]}]': finished 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:47.727212+0000 mon.c (mon.1) 15 : cluster [INF] mon.c calling monitor election 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:47.727212+0000 mon.c (mon.1) 15 : cluster [INF] mon.c calling monitor election 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: audit 2026-03-10T10:18:47.728739+0000 mon.a (mon.0) 618 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: audit 2026-03-10T10:18:47.728739+0000 mon.a (mon.0) 618 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: audit 2026-03-10T10:18:47.728811+0000 mon.a (mon.0) 619 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: audit 2026-03-10T10:18:47.728811+0000 mon.a (mon.0) 619 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: audit 2026-03-10T10:18:47.728887+0000 mon.a (mon.0) 620 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: audit 2026-03-10T10:18:47.728887+0000 mon.a (mon.0) 620 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:47.729207+0000 mon.a (mon.0) 621 : cluster [INF] mon.a calling monitor election 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:47.729207+0000 mon.a (mon.0) 621 : cluster [INF] mon.a calling monitor election 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:47.731049+0000 mon.b (mon.2) 10 : cluster [INF] mon.b calling monitor election 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:47.731049+0000 mon.b (mon.2) 10 : cluster [INF] mon.b calling monitor election 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:47.787173+0000 mgr.a (mgr.14150) 189 : cluster [DBG] pgmap v123: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:47.787173+0000 mgr.a (mgr.14150) 189 : cluster [DBG] pgmap v123: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:49.787439+0000 mgr.a (mgr.14150) 190 : cluster [DBG] pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:49.787439+0000 mgr.a (mgr.14150) 190 : cluster [DBG] pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:51.787807+0000 mgr.a (mgr.14150) 191 : cluster [DBG] pgmap v125: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:51.787807+0000 mgr.a (mgr.14150) 191 : cluster [DBG] pgmap v125: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:52.830012+0000 mon.a (mon.0) 622 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:52.830012+0000 mon.a (mon.0) 622 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T10:18:53.760 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.060857+0000 mon.a (mon.0) 623 : cluster [INF] overall HEALTH_OK 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.060857+0000 mon.a (mon.0) 623 : cluster [INF] overall HEALTH_OK 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.060947+0000 mon.a (mon.0) 624 : cluster [INF] mon.a calling monitor election 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.060947+0000 mon.a (mon.0) 624 : cluster [INF] mon.a calling monitor election 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.135623+0000 mon.c (mon.1) 16 : cluster [INF] mon.c calling monitor election 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.135623+0000 mon.c (mon.1) 16 : cluster [INF] mon.c calling monitor election 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.223820+0000 mon.a (mon.0) 625 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.223820+0000 mon.a (mon.0) 625 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307561+0000 mon.a (mon.0) 626 : cluster [DBG] monmap epoch 4 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307561+0000 mon.a (mon.0) 626 : cluster [DBG] monmap epoch 4 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307571+0000 mon.a (mon.0) 627 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307571+0000 mon.a (mon.0) 627 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307575+0000 mon.a (mon.0) 628 : cluster [DBG] last_changed 2026-03-10T10:18:47.676101+0000 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307575+0000 mon.a (mon.0) 628 : cluster [DBG] last_changed 2026-03-10T10:18:47.676101+0000 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307579+0000 mon.a (mon.0) 629 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307579+0000 mon.a (mon.0) 629 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307582+0000 mon.a (mon.0) 630 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307582+0000 mon.a (mon.0) 630 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307585+0000 mon.a (mon.0) 631 : cluster [DBG] election_strategy: 1 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307585+0000 mon.a (mon.0) 631 : cluster [DBG] election_strategy: 1 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307590+0000 mon.a (mon.0) 632 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a; crush_location {datacenter=a} 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307590+0000 mon.a (mon.0) 632 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a; crush_location {datacenter=a} 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307593+0000 mon.a (mon.0) 633 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307593+0000 mon.a (mon.0) 633 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307597+0000 mon.a (mon.0) 634 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307597+0000 mon.a (mon.0) 634 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307913+0000 mon.a (mon.0) 635 : cluster [DBG] fsmap 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307913+0000 mon.a (mon.0) 635 : cluster [DBG] fsmap 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307925+0000 mon.a (mon.0) 636 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.307925+0000 mon.a (mon.0) 636 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.308202+0000 mon.a (mon.0) 637 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.308202+0000 mon.a (mon.0) 637 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.308276+0000 mon.a (mon.0) 638 : cluster [INF] overall HEALTH_OK 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: cluster 2026-03-10T10:18:53.308276+0000 mon.a (mon.0) 638 : cluster [INF] overall HEALTH_OK 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: audit 2026-03-10T10:18:53.331906+0000 mon.a (mon.0) 639 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon set_location", "name": "b", "args": ["datacenter=b", "rack=2"]}]: dispatch 2026-03-10T10:18:53.761 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:18:53 vm09 bash[20916]: audit 2026-03-10T10:18:53.331906+0000 mon.a (mon.0) 639 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon set_location", "name": "b", "args": ["datacenter=b", "rack=2"]}]: dispatch 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: audit 2026-03-10T10:18:46.922198+0000 mgr.a (mgr.14150) 184 : audit [DBG] from='client.14385 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: audit 2026-03-10T10:18:46.922198+0000 mgr.a (mgr.14150) 184 : audit [DBG] from='client.14385 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: audit 2026-03-10T10:18:47.090343+0000 mgr.a (mgr.14150) 185 : audit [DBG] from='client.24268 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: audit 2026-03-10T10:18:47.090343+0000 mgr.a (mgr.14150) 185 : audit [DBG] from='client.24268 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: audit 2026-03-10T10:18:47.263971+0000 mgr.a (mgr.14150) 186 : audit [DBG] from='client.14397 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: audit 2026-03-10T10:18:47.263971+0000 mgr.a (mgr.14150) 186 : audit [DBG] from='client.14397 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cephadm 2026-03-10T10:18:47.265525+0000 mgr.a (mgr.14150) 187 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cephadm 2026-03-10T10:18:47.265525+0000 mgr.a (mgr.14150) 187 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cephadm 2026-03-10T10:18:47.675866+0000 mgr.a (mgr.14150) 188 : cephadm [INF] Setting crush location for mon a to {datacenter=a} 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cephadm 2026-03-10T10:18:47.675866+0000 mgr.a (mgr.14150) 188 : cephadm [INF] Setting crush location for mon a to {datacenter=a} 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: audit 2026-03-10T10:18:47.726201+0000 mon.a (mon.0) 617 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "a", "args": ["datacenter=a"]}]': finished 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: audit 2026-03-10T10:18:47.726201+0000 mon.a (mon.0) 617 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "a", "args": ["datacenter=a"]}]': finished 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:47.727212+0000 mon.c (mon.1) 15 : cluster [INF] mon.c calling monitor election 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:47.727212+0000 mon.c (mon.1) 15 : cluster [INF] mon.c calling monitor election 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: audit 2026-03-10T10:18:47.728739+0000 mon.a (mon.0) 618 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: audit 2026-03-10T10:18:47.728739+0000 mon.a (mon.0) 618 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: audit 2026-03-10T10:18:47.728811+0000 mon.a (mon.0) 619 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: audit 2026-03-10T10:18:47.728811+0000 mon.a (mon.0) 619 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: audit 2026-03-10T10:18:47.728887+0000 mon.a (mon.0) 620 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: audit 2026-03-10T10:18:47.728887+0000 mon.a (mon.0) 620 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:47.729207+0000 mon.a (mon.0) 621 : cluster [INF] mon.a calling monitor election 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:47.729207+0000 mon.a (mon.0) 621 : cluster [INF] mon.a calling monitor election 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:47.731049+0000 mon.b (mon.2) 10 : cluster [INF] mon.b calling monitor election 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:47.731049+0000 mon.b (mon.2) 10 : cluster [INF] mon.b calling monitor election 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:47.787173+0000 mgr.a (mgr.14150) 189 : cluster [DBG] pgmap v123: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:47.787173+0000 mgr.a (mgr.14150) 189 : cluster [DBG] pgmap v123: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:49.787439+0000 mgr.a (mgr.14150) 190 : cluster [DBG] pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:49.787439+0000 mgr.a (mgr.14150) 190 : cluster [DBG] pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:51.787807+0000 mgr.a (mgr.14150) 191 : cluster [DBG] pgmap v125: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:51.787807+0000 mgr.a (mgr.14150) 191 : cluster [DBG] pgmap v125: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:52.830012+0000 mon.a (mon.0) 622 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:52.830012+0000 mon.a (mon.0) 622 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.060857+0000 mon.a (mon.0) 623 : cluster [INF] overall HEALTH_OK 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.060857+0000 mon.a (mon.0) 623 : cluster [INF] overall HEALTH_OK 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.060947+0000 mon.a (mon.0) 624 : cluster [INF] mon.a calling monitor election 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.060947+0000 mon.a (mon.0) 624 : cluster [INF] mon.a calling monitor election 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.135623+0000 mon.c (mon.1) 16 : cluster [INF] mon.c calling monitor election 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.135623+0000 mon.c (mon.1) 16 : cluster [INF] mon.c calling monitor election 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.223820+0000 mon.a (mon.0) 625 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.223820+0000 mon.a (mon.0) 625 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307561+0000 mon.a (mon.0) 626 : cluster [DBG] monmap epoch 4 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307561+0000 mon.a (mon.0) 626 : cluster [DBG] monmap epoch 4 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307571+0000 mon.a (mon.0) 627 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307571+0000 mon.a (mon.0) 627 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307575+0000 mon.a (mon.0) 628 : cluster [DBG] last_changed 2026-03-10T10:18:47.676101+0000 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307575+0000 mon.a (mon.0) 628 : cluster [DBG] last_changed 2026-03-10T10:18:47.676101+0000 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307579+0000 mon.a (mon.0) 629 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307579+0000 mon.a (mon.0) 629 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307582+0000 mon.a (mon.0) 630 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307582+0000 mon.a (mon.0) 630 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307585+0000 mon.a (mon.0) 631 : cluster [DBG] election_strategy: 1 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307585+0000 mon.a (mon.0) 631 : cluster [DBG] election_strategy: 1 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307590+0000 mon.a (mon.0) 632 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a; crush_location {datacenter=a} 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307590+0000 mon.a (mon.0) 632 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a; crush_location {datacenter=a} 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307593+0000 mon.a (mon.0) 633 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307593+0000 mon.a (mon.0) 633 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307597+0000 mon.a (mon.0) 634 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307597+0000 mon.a (mon.0) 634 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307913+0000 mon.a (mon.0) 635 : cluster [DBG] fsmap 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307913+0000 mon.a (mon.0) 635 : cluster [DBG] fsmap 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307925+0000 mon.a (mon.0) 636 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.307925+0000 mon.a (mon.0) 636 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.308202+0000 mon.a (mon.0) 637 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.308202+0000 mon.a (mon.0) 637 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.308276+0000 mon.a (mon.0) 638 : cluster [INF] overall HEALTH_OK 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: cluster 2026-03-10T10:18:53.308276+0000 mon.a (mon.0) 638 : cluster [INF] overall HEALTH_OK 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: audit 2026-03-10T10:18:53.331906+0000 mon.a (mon.0) 639 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon set_location", "name": "b", "args": ["datacenter=b", "rack=2"]}]: dispatch 2026-03-10T10:18:53.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:18:53 vm03 bash[21274]: audit 2026-03-10T10:18:53.331906+0000 mon.a (mon.0) 639 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon set_location", "name": "b", "args": ["datacenter=b", "rack=2"]}]: dispatch 2026-03-10T10:18:53.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: audit 2026-03-10T10:18:46.922198+0000 mgr.a (mgr.14150) 184 : audit [DBG] from='client.14385 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:53.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: audit 2026-03-10T10:18:46.922198+0000 mgr.a (mgr.14150) 184 : audit [DBG] from='client.14385 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: audit 2026-03-10T10:18:47.090343+0000 mgr.a (mgr.14150) 185 : audit [DBG] from='client.24268 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: audit 2026-03-10T10:18:47.090343+0000 mgr.a (mgr.14150) 185 : audit [DBG] from='client.24268 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: audit 2026-03-10T10:18:47.263971+0000 mgr.a (mgr.14150) 186 : audit [DBG] from='client.14397 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: audit 2026-03-10T10:18:47.263971+0000 mgr.a (mgr.14150) 186 : audit [DBG] from='client.14397 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cephadm 2026-03-10T10:18:47.265525+0000 mgr.a (mgr.14150) 187 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cephadm 2026-03-10T10:18:47.265525+0000 mgr.a (mgr.14150) 187 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cephadm 2026-03-10T10:18:47.675866+0000 mgr.a (mgr.14150) 188 : cephadm [INF] Setting crush location for mon a to {datacenter=a} 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cephadm 2026-03-10T10:18:47.675866+0000 mgr.a (mgr.14150) 188 : cephadm [INF] Setting crush location for mon a to {datacenter=a} 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: audit 2026-03-10T10:18:47.726201+0000 mon.a (mon.0) 617 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "a", "args": ["datacenter=a"]}]': finished 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: audit 2026-03-10T10:18:47.726201+0000 mon.a (mon.0) 617 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "a", "args": ["datacenter=a"]}]': finished 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:47.727212+0000 mon.c (mon.1) 15 : cluster [INF] mon.c calling monitor election 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:47.727212+0000 mon.c (mon.1) 15 : cluster [INF] mon.c calling monitor election 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: audit 2026-03-10T10:18:47.728739+0000 mon.a (mon.0) 618 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: audit 2026-03-10T10:18:47.728739+0000 mon.a (mon.0) 618 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: audit 2026-03-10T10:18:47.728811+0000 mon.a (mon.0) 619 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: audit 2026-03-10T10:18:47.728811+0000 mon.a (mon.0) 619 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: audit 2026-03-10T10:18:47.728887+0000 mon.a (mon.0) 620 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: audit 2026-03-10T10:18:47.728887+0000 mon.a (mon.0) 620 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:47.729207+0000 mon.a (mon.0) 621 : cluster [INF] mon.a calling monitor election 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:47.729207+0000 mon.a (mon.0) 621 : cluster [INF] mon.a calling monitor election 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:47.731049+0000 mon.b (mon.2) 10 : cluster [INF] mon.b calling monitor election 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:47.731049+0000 mon.b (mon.2) 10 : cluster [INF] mon.b calling monitor election 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:47.787173+0000 mgr.a (mgr.14150) 189 : cluster [DBG] pgmap v123: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:47.787173+0000 mgr.a (mgr.14150) 189 : cluster [DBG] pgmap v123: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:49.787439+0000 mgr.a (mgr.14150) 190 : cluster [DBG] pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:49.787439+0000 mgr.a (mgr.14150) 190 : cluster [DBG] pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:51.787807+0000 mgr.a (mgr.14150) 191 : cluster [DBG] pgmap v125: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:51.787807+0000 mgr.a (mgr.14150) 191 : cluster [DBG] pgmap v125: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:52.830012+0000 mon.a (mon.0) 622 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:52.830012+0000 mon.a (mon.0) 622 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.060857+0000 mon.a (mon.0) 623 : cluster [INF] overall HEALTH_OK 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.060857+0000 mon.a (mon.0) 623 : cluster [INF] overall HEALTH_OK 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.060947+0000 mon.a (mon.0) 624 : cluster [INF] mon.a calling monitor election 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.060947+0000 mon.a (mon.0) 624 : cluster [INF] mon.a calling monitor election 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.135623+0000 mon.c (mon.1) 16 : cluster [INF] mon.c calling monitor election 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.135623+0000 mon.c (mon.1) 16 : cluster [INF] mon.c calling monitor election 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.223820+0000 mon.a (mon.0) 625 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.223820+0000 mon.a (mon.0) 625 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307561+0000 mon.a (mon.0) 626 : cluster [DBG] monmap epoch 4 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307561+0000 mon.a (mon.0) 626 : cluster [DBG] monmap epoch 4 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307571+0000 mon.a (mon.0) 627 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307571+0000 mon.a (mon.0) 627 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307575+0000 mon.a (mon.0) 628 : cluster [DBG] last_changed 2026-03-10T10:18:47.676101+0000 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307575+0000 mon.a (mon.0) 628 : cluster [DBG] last_changed 2026-03-10T10:18:47.676101+0000 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307579+0000 mon.a (mon.0) 629 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307579+0000 mon.a (mon.0) 629 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307582+0000 mon.a (mon.0) 630 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:18:53.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307582+0000 mon.a (mon.0) 630 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307585+0000 mon.a (mon.0) 631 : cluster [DBG] election_strategy: 1 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307585+0000 mon.a (mon.0) 631 : cluster [DBG] election_strategy: 1 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307590+0000 mon.a (mon.0) 632 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a; crush_location {datacenter=a} 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307590+0000 mon.a (mon.0) 632 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a; crush_location {datacenter=a} 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307593+0000 mon.a (mon.0) 633 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307593+0000 mon.a (mon.0) 633 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307597+0000 mon.a (mon.0) 634 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307597+0000 mon.a (mon.0) 634 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307913+0000 mon.a (mon.0) 635 : cluster [DBG] fsmap 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307913+0000 mon.a (mon.0) 635 : cluster [DBG] fsmap 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307925+0000 mon.a (mon.0) 636 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.307925+0000 mon.a (mon.0) 636 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.308202+0000 mon.a (mon.0) 637 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.308202+0000 mon.a (mon.0) 637 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.308276+0000 mon.a (mon.0) 638 : cluster [INF] overall HEALTH_OK 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: cluster 2026-03-10T10:18:53.308276+0000 mon.a (mon.0) 638 : cluster [INF] overall HEALTH_OK 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: audit 2026-03-10T10:18:53.331906+0000 mon.a (mon.0) 639 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon set_location", "name": "b", "args": ["datacenter=b", "rack=2"]}]: dispatch 2026-03-10T10:18:53.918 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:18:53 vm00 bash[20709]: audit 2026-03-10T10:18:53.331906+0000 mon.a (mon.0) 639 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon set_location", "name": "b", "args": ["datacenter=b", "rack=2"]}]: dispatch 2026-03-10T10:19:03.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:18:57.788710+0000 mgr.a (mgr.14150) 195 : cluster [DBG] pgmap v128: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:03.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:18:57.788710+0000 mgr.a (mgr.14150) 195 : cluster [DBG] pgmap v128: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:03.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cephadm 2026-03-10T10:18:58.465995+0000 mgr.a (mgr.14150) 196 : cephadm [INF] Setting crush location for mon c to {datacenter=a,rack=3} 2026-03-10T10:19:03.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cephadm 2026-03-10T10:18:58.465995+0000 mgr.a (mgr.14150) 196 : cephadm [INF] Setting crush location for mon c to {datacenter=a,rack=3} 2026-03-10T10:19:03.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: audit 2026-03-10T10:18:58.546490+0000 mon.a (mon.0) 664 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "c", "args": ["datacenter=a", "rack=3"]}]': finished 2026-03-10T10:19:03.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: audit 2026-03-10T10:18:58.546490+0000 mon.a (mon.0) 664 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "c", "args": ["datacenter=a", "rack=3"]}]': finished 2026-03-10T10:19:03.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:18:58.548509+0000 mon.c (mon.1) 19 : cluster [INF] mon.c calling monitor election 2026-03-10T10:19:03.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:18:58.548509+0000 mon.c (mon.1) 19 : cluster [INF] mon.c calling monitor election 2026-03-10T10:19:03.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: audit 2026-03-10T10:18:58.548710+0000 mon.a (mon.0) 665 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:03.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: audit 2026-03-10T10:18:58.548710+0000 mon.a (mon.0) 665 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:03.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:18:58.548997+0000 mon.b (mon.2) 12 : cluster [INF] mon.b calling monitor election 2026-03-10T10:19:03.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:18:58.548997+0000 mon.b (mon.2) 12 : cluster [INF] mon.b calling monitor election 2026-03-10T10:19:03.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:18:58.549657+0000 mon.a (mon.0) 666 : cluster [INF] mon.a calling monitor election 2026-03-10T10:19:03.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:18:58.549657+0000 mon.a (mon.0) 666 : cluster [INF] mon.a calling monitor election 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: audit 2026-03-10T10:18:58.550794+0000 mon.a (mon.0) 667 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: audit 2026-03-10T10:18:58.550794+0000 mon.a (mon.0) 667 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: audit 2026-03-10T10:18:58.550867+0000 mon.a (mon.0) 668 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: audit 2026-03-10T10:18:58.550867+0000 mon.a (mon.0) 668 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: audit 2026-03-10T10:18:58.550903+0000 mon.a (mon.0) 669 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: audit 2026-03-10T10:18:58.550903+0000 mon.a (mon.0) 669 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:18:59.788985+0000 mgr.a (mgr.14150) 197 : cluster [DBG] pgmap v129: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:18:59.788985+0000 mgr.a (mgr.14150) 197 : cluster [DBG] pgmap v129: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:01.789234+0000 mgr.a (mgr.14150) 198 : cluster [DBG] pgmap v130: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:01.789234+0000 mgr.a (mgr.14150) 198 : cluster [DBG] pgmap v130: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.567527+0000 mon.a (mon.0) 670 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,2) 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.567527+0000 mon.a (mon.0) 670 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,2) 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.569044+0000 mon.a (mon.0) 671 : cluster [INF] overall HEALTH_OK 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.569044+0000 mon.a (mon.0) 671 : cluster [INF] overall HEALTH_OK 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.569072+0000 mon.a (mon.0) 672 : cluster [INF] mon.a calling monitor election 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.569072+0000 mon.a (mon.0) 672 : cluster [INF] mon.a calling monitor election 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.571237+0000 mon.a (mon.0) 673 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.571237+0000 mon.a (mon.0) 673 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575669+0000 mon.a (mon.0) 674 : cluster [DBG] monmap epoch 6 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575669+0000 mon.a (mon.0) 674 : cluster [DBG] monmap epoch 6 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575682+0000 mon.a (mon.0) 675 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575682+0000 mon.a (mon.0) 675 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575686+0000 mon.a (mon.0) 676 : cluster [DBG] last_changed 2026-03-10T10:18:58.467065+0000 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575686+0000 mon.a (mon.0) 676 : cluster [DBG] last_changed 2026-03-10T10:18:58.467065+0000 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575693+0000 mon.a (mon.0) 677 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575693+0000 mon.a (mon.0) 677 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575697+0000 mon.a (mon.0) 678 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575697+0000 mon.a (mon.0) 678 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575702+0000 mon.a (mon.0) 679 : cluster [DBG] election_strategy: 1 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575702+0000 mon.a (mon.0) 679 : cluster [DBG] election_strategy: 1 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575708+0000 mon.a (mon.0) 680 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a; crush_location {datacenter=a} 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575708+0000 mon.a (mon.0) 680 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a; crush_location {datacenter=a} 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575714+0000 mon.a (mon.0) 681 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575714+0000 mon.a (mon.0) 681 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575720+0000 mon.a (mon.0) 682 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.575720+0000 mon.a (mon.0) 682 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.576278+0000 mon.a (mon.0) 683 : cluster [DBG] fsmap 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.576278+0000 mon.a (mon.0) 683 : cluster [DBG] fsmap 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.576300+0000 mon.a (mon.0) 684 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.576300+0000 mon.a (mon.0) 684 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.576819+0000 mon.a (mon.0) 685 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.576819+0000 mon.a (mon.0) 685 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.576939+0000 mon.a (mon.0) 686 : cluster [INF] overall HEALTH_OK 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: cluster 2026-03-10T10:19:03.576939+0000 mon.a (mon.0) 686 : cluster [INF] overall HEALTH_OK 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: audit 2026-03-10T10:19:03.601372+0000 mon.a (mon.0) 687 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: audit 2026-03-10T10:19:03.601372+0000 mon.a (mon.0) 687 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: audit 2026-03-10T10:19:03.608318+0000 mon.a (mon.0) 688 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:03.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:03 vm00 bash[20709]: audit 2026-03-10T10:19:03.608318+0000 mon.a (mon.0) 688 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:18:57.788710+0000 mgr.a (mgr.14150) 195 : cluster [DBG] pgmap v128: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:18:57.788710+0000 mgr.a (mgr.14150) 195 : cluster [DBG] pgmap v128: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cephadm 2026-03-10T10:18:58.465995+0000 mgr.a (mgr.14150) 196 : cephadm [INF] Setting crush location for mon c to {datacenter=a,rack=3} 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cephadm 2026-03-10T10:18:58.465995+0000 mgr.a (mgr.14150) 196 : cephadm [INF] Setting crush location for mon c to {datacenter=a,rack=3} 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: audit 2026-03-10T10:18:58.546490+0000 mon.a (mon.0) 664 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "c", "args": ["datacenter=a", "rack=3"]}]': finished 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: audit 2026-03-10T10:18:58.546490+0000 mon.a (mon.0) 664 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "c", "args": ["datacenter=a", "rack=3"]}]': finished 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:18:58.548509+0000 mon.c (mon.1) 19 : cluster [INF] mon.c calling monitor election 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:18:58.548509+0000 mon.c (mon.1) 19 : cluster [INF] mon.c calling monitor election 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: audit 2026-03-10T10:18:58.548710+0000 mon.a (mon.0) 665 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: audit 2026-03-10T10:18:58.548710+0000 mon.a (mon.0) 665 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:18:58.548997+0000 mon.b (mon.2) 12 : cluster [INF] mon.b calling monitor election 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:18:58.548997+0000 mon.b (mon.2) 12 : cluster [INF] mon.b calling monitor election 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:18:58.549657+0000 mon.a (mon.0) 666 : cluster [INF] mon.a calling monitor election 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:18:58.549657+0000 mon.a (mon.0) 666 : cluster [INF] mon.a calling monitor election 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: audit 2026-03-10T10:18:58.550794+0000 mon.a (mon.0) 667 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: audit 2026-03-10T10:18:58.550794+0000 mon.a (mon.0) 667 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: audit 2026-03-10T10:18:58.550867+0000 mon.a (mon.0) 668 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: audit 2026-03-10T10:18:58.550867+0000 mon.a (mon.0) 668 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: audit 2026-03-10T10:18:58.550903+0000 mon.a (mon.0) 669 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: audit 2026-03-10T10:18:58.550903+0000 mon.a (mon.0) 669 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:18:59.788985+0000 mgr.a (mgr.14150) 197 : cluster [DBG] pgmap v129: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:18:59.788985+0000 mgr.a (mgr.14150) 197 : cluster [DBG] pgmap v129: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:01.789234+0000 mgr.a (mgr.14150) 198 : cluster [DBG] pgmap v130: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:01.789234+0000 mgr.a (mgr.14150) 198 : cluster [DBG] pgmap v130: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.567527+0000 mon.a (mon.0) 670 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,2) 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.567527+0000 mon.a (mon.0) 670 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,2) 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.569044+0000 mon.a (mon.0) 671 : cluster [INF] overall HEALTH_OK 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.569044+0000 mon.a (mon.0) 671 : cluster [INF] overall HEALTH_OK 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.569072+0000 mon.a (mon.0) 672 : cluster [INF] mon.a calling monitor election 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.569072+0000 mon.a (mon.0) 672 : cluster [INF] mon.a calling monitor election 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.571237+0000 mon.a (mon.0) 673 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.571237+0000 mon.a (mon.0) 673 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575669+0000 mon.a (mon.0) 674 : cluster [DBG] monmap epoch 6 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575669+0000 mon.a (mon.0) 674 : cluster [DBG] monmap epoch 6 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575682+0000 mon.a (mon.0) 675 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:19:04.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575682+0000 mon.a (mon.0) 675 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575686+0000 mon.a (mon.0) 676 : cluster [DBG] last_changed 2026-03-10T10:18:58.467065+0000 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575686+0000 mon.a (mon.0) 676 : cluster [DBG] last_changed 2026-03-10T10:18:58.467065+0000 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575693+0000 mon.a (mon.0) 677 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575693+0000 mon.a (mon.0) 677 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575697+0000 mon.a (mon.0) 678 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575697+0000 mon.a (mon.0) 678 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575702+0000 mon.a (mon.0) 679 : cluster [DBG] election_strategy: 1 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575702+0000 mon.a (mon.0) 679 : cluster [DBG] election_strategy: 1 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575708+0000 mon.a (mon.0) 680 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a; crush_location {datacenter=a} 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575708+0000 mon.a (mon.0) 680 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a; crush_location {datacenter=a} 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575714+0000 mon.a (mon.0) 681 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575714+0000 mon.a (mon.0) 681 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575720+0000 mon.a (mon.0) 682 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.575720+0000 mon.a (mon.0) 682 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.576278+0000 mon.a (mon.0) 683 : cluster [DBG] fsmap 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.576278+0000 mon.a (mon.0) 683 : cluster [DBG] fsmap 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.576300+0000 mon.a (mon.0) 684 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.576300+0000 mon.a (mon.0) 684 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.576819+0000 mon.a (mon.0) 685 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.576819+0000 mon.a (mon.0) 685 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.576939+0000 mon.a (mon.0) 686 : cluster [INF] overall HEALTH_OK 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: cluster 2026-03-10T10:19:03.576939+0000 mon.a (mon.0) 686 : cluster [INF] overall HEALTH_OK 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: audit 2026-03-10T10:19:03.601372+0000 mon.a (mon.0) 687 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: audit 2026-03-10T10:19:03.601372+0000 mon.a (mon.0) 687 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: audit 2026-03-10T10:19:03.608318+0000 mon.a (mon.0) 688 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.011 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:03 vm09 bash[20916]: audit 2026-03-10T10:19:03.608318+0000 mon.a (mon.0) 688 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:18:57.788710+0000 mgr.a (mgr.14150) 195 : cluster [DBG] pgmap v128: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:18:57.788710+0000 mgr.a (mgr.14150) 195 : cluster [DBG] pgmap v128: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cephadm 2026-03-10T10:18:58.465995+0000 mgr.a (mgr.14150) 196 : cephadm [INF] Setting crush location for mon c to {datacenter=a,rack=3} 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cephadm 2026-03-10T10:18:58.465995+0000 mgr.a (mgr.14150) 196 : cephadm [INF] Setting crush location for mon c to {datacenter=a,rack=3} 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: audit 2026-03-10T10:18:58.546490+0000 mon.a (mon.0) 664 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "c", "args": ["datacenter=a", "rack=3"]}]': finished 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: audit 2026-03-10T10:18:58.546490+0000 mon.a (mon.0) 664 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd='[{"prefix": "mon set_location", "name": "c", "args": ["datacenter=a", "rack=3"]}]': finished 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:18:58.548509+0000 mon.c (mon.1) 19 : cluster [INF] mon.c calling monitor election 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:18:58.548509+0000 mon.c (mon.1) 19 : cluster [INF] mon.c calling monitor election 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: audit 2026-03-10T10:18:58.548710+0000 mon.a (mon.0) 665 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: audit 2026-03-10T10:18:58.548710+0000 mon.a (mon.0) 665 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:18:58.548997+0000 mon.b (mon.2) 12 : cluster [INF] mon.b calling monitor election 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:18:58.548997+0000 mon.b (mon.2) 12 : cluster [INF] mon.b calling monitor election 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:18:58.549657+0000 mon.a (mon.0) 666 : cluster [INF] mon.a calling monitor election 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:18:58.549657+0000 mon.a (mon.0) 666 : cluster [INF] mon.a calling monitor election 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: audit 2026-03-10T10:18:58.550794+0000 mon.a (mon.0) 667 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: audit 2026-03-10T10:18:58.550794+0000 mon.a (mon.0) 667 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: audit 2026-03-10T10:18:58.550867+0000 mon.a (mon.0) 668 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: audit 2026-03-10T10:18:58.550867+0000 mon.a (mon.0) 668 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: audit 2026-03-10T10:18:58.550903+0000 mon.a (mon.0) 669 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: audit 2026-03-10T10:18:58.550903+0000 mon.a (mon.0) 669 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:18:59.788985+0000 mgr.a (mgr.14150) 197 : cluster [DBG] pgmap v129: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:18:59.788985+0000 mgr.a (mgr.14150) 197 : cluster [DBG] pgmap v129: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:01.789234+0000 mgr.a (mgr.14150) 198 : cluster [DBG] pgmap v130: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:01.789234+0000 mgr.a (mgr.14150) 198 : cluster [DBG] pgmap v130: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:04.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.567527+0000 mon.a (mon.0) 670 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,2) 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.567527+0000 mon.a (mon.0) 670 : cluster [INF] mon.a is new leader, mons a,b in quorum (ranks 0,2) 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.569044+0000 mon.a (mon.0) 671 : cluster [INF] overall HEALTH_OK 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.569044+0000 mon.a (mon.0) 671 : cluster [INF] overall HEALTH_OK 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.569072+0000 mon.a (mon.0) 672 : cluster [INF] mon.a calling monitor election 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.569072+0000 mon.a (mon.0) 672 : cluster [INF] mon.a calling monitor election 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.571237+0000 mon.a (mon.0) 673 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.571237+0000 mon.a (mon.0) 673 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575669+0000 mon.a (mon.0) 674 : cluster [DBG] monmap epoch 6 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575669+0000 mon.a (mon.0) 674 : cluster [DBG] monmap epoch 6 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575682+0000 mon.a (mon.0) 675 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575682+0000 mon.a (mon.0) 675 : cluster [DBG] fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575686+0000 mon.a (mon.0) 676 : cluster [DBG] last_changed 2026-03-10T10:18:58.467065+0000 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575686+0000 mon.a (mon.0) 676 : cluster [DBG] last_changed 2026-03-10T10:18:58.467065+0000 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575693+0000 mon.a (mon.0) 677 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575693+0000 mon.a (mon.0) 677 : cluster [DBG] created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575697+0000 mon.a (mon.0) 678 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575697+0000 mon.a (mon.0) 678 : cluster [DBG] min_mon_release 19 (squid) 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575702+0000 mon.a (mon.0) 679 : cluster [DBG] election_strategy: 1 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575702+0000 mon.a (mon.0) 679 : cluster [DBG] election_strategy: 1 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575708+0000 mon.a (mon.0) 680 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a; crush_location {datacenter=a} 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575708+0000 mon.a (mon.0) 680 : cluster [DBG] 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a; crush_location {datacenter=a} 2026-03-10T10:19:04.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575714+0000 mon.a (mon.0) 681 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-10T10:19:04.083 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575714+0000 mon.a (mon.0) 681 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-10T10:19:04.083 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575720+0000 mon.a (mon.0) 682 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-10T10:19:04.083 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.575720+0000 mon.a (mon.0) 682 : cluster [DBG] 2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-10T10:19:04.083 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.576278+0000 mon.a (mon.0) 683 : cluster [DBG] fsmap 2026-03-10T10:19:04.083 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.576278+0000 mon.a (mon.0) 683 : cluster [DBG] fsmap 2026-03-10T10:19:04.083 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.576300+0000 mon.a (mon.0) 684 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:19:04.083 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.576300+0000 mon.a (mon.0) 684 : cluster [DBG] osdmap e22: 3 total, 3 up, 3 in 2026-03-10T10:19:04.083 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.576819+0000 mon.a (mon.0) 685 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-10T10:19:04.083 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.576819+0000 mon.a (mon.0) 685 : cluster [DBG] mgrmap e14: a(active, since 3m), standbys: b 2026-03-10T10:19:04.083 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.576939+0000 mon.a (mon.0) 686 : cluster [INF] overall HEALTH_OK 2026-03-10T10:19:04.084 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: cluster 2026-03-10T10:19:03.576939+0000 mon.a (mon.0) 686 : cluster [INF] overall HEALTH_OK 2026-03-10T10:19:04.084 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: audit 2026-03-10T10:19:03.601372+0000 mon.a (mon.0) 687 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.084 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: audit 2026-03-10T10:19:03.601372+0000 mon.a (mon.0) 687 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.084 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: audit 2026-03-10T10:19:03.608318+0000 mon.a (mon.0) 688 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.084 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:03 vm03 bash[21274]: audit 2026-03-10T10:19:03.608318+0000 mon.a (mon.0) 688 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.632824+0000 mon.a (mon.0) 689 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.632824+0000 mon.a (mon.0) 689 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.643321+0000 mon.a (mon.0) 690 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.643321+0000 mon.a (mon.0) 690 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: cephadm 2026-03-10T10:19:03.661636+0000 mgr.a (mgr.14150) 199 : cephadm [INF] Reconfiguring mgr.a (monmap changed)... 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: cephadm 2026-03-10T10:19:03.661636+0000 mgr.a (mgr.14150) 199 : cephadm [INF] Reconfiguring mgr.a (monmap changed)... 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.664208+0000 mon.a (mon.0) 691 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.664208+0000 mon.a (mon.0) 691 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.665056+0000 mon.a (mon.0) 692 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.665056+0000 mon.a (mon.0) 692 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.665594+0000 mon.a (mon.0) 693 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.665594+0000 mon.a (mon.0) 693 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: cephadm 2026-03-10T10:19:03.666536+0000 mgr.a (mgr.14150) 200 : cephadm [INF] Reconfiguring daemon mgr.a on vm00 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: cephadm 2026-03-10T10:19:03.666536+0000 mgr.a (mgr.14150) 200 : cephadm [INF] Reconfiguring daemon mgr.a on vm00 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.669658+0000 mon.a (mon.0) 694 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.669658+0000 mon.a (mon.0) 694 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.680779+0000 mon.a (mon.0) 695 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.680779+0000 mon.a (mon.0) 695 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.686238+0000 mon.a (mon.0) 696 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.686238+0000 mon.a (mon.0) 696 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.695126+0000 mon.a (mon.0) 697 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:03.695126+0000 mon.a (mon.0) 697 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:04.228112+0000 mon.a (mon.0) 698 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:04.228112+0000 mon.a (mon.0) 698 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:04.233794+0000 mon.a (mon.0) 699 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:04.233794+0000 mon.a (mon.0) 699 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:04.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:04.235007+0000 mon.a (mon.0) 700 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T10:19:04.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:04.235007+0000 mon.a (mon.0) 700 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T10:19:04.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:04.235469+0000 mon.a (mon.0) 701 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:04.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:04 vm00 bash[20709]: audit 2026-03-10T10:19:04.235469+0000 mon.a (mon.0) 701 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.632824+0000 mon.a (mon.0) 689 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.632824+0000 mon.a (mon.0) 689 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.643321+0000 mon.a (mon.0) 690 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.643321+0000 mon.a (mon.0) 690 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: cephadm 2026-03-10T10:19:03.661636+0000 mgr.a (mgr.14150) 199 : cephadm [INF] Reconfiguring mgr.a (monmap changed)... 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: cephadm 2026-03-10T10:19:03.661636+0000 mgr.a (mgr.14150) 199 : cephadm [INF] Reconfiguring mgr.a (monmap changed)... 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.664208+0000 mon.a (mon.0) 691 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.664208+0000 mon.a (mon.0) 691 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.665056+0000 mon.a (mon.0) 692 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.665056+0000 mon.a (mon.0) 692 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.665594+0000 mon.a (mon.0) 693 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.665594+0000 mon.a (mon.0) 693 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: cephadm 2026-03-10T10:19:03.666536+0000 mgr.a (mgr.14150) 200 : cephadm [INF] Reconfiguring daemon mgr.a on vm00 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: cephadm 2026-03-10T10:19:03.666536+0000 mgr.a (mgr.14150) 200 : cephadm [INF] Reconfiguring daemon mgr.a on vm00 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.669658+0000 mon.a (mon.0) 694 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.669658+0000 mon.a (mon.0) 694 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.680779+0000 mon.a (mon.0) 695 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.680779+0000 mon.a (mon.0) 695 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.686238+0000 mon.a (mon.0) 696 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.686238+0000 mon.a (mon.0) 696 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.695126+0000 mon.a (mon.0) 697 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:03.695126+0000 mon.a (mon.0) 697 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:04.228112+0000 mon.a (mon.0) 698 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:04.228112+0000 mon.a (mon.0) 698 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:04.233794+0000 mon.a (mon.0) 699 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:04.233794+0000 mon.a (mon.0) 699 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:04.235007+0000 mon.a (mon.0) 700 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T10:19:05.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:04.235007+0000 mon.a (mon.0) 700 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T10:19:05.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:04.235469+0000 mon.a (mon.0) 701 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:05.010 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:04 vm09 bash[20916]: audit 2026-03-10T10:19:04.235469+0000 mon.a (mon.0) 701 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.632824+0000 mon.a (mon.0) 689 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.632824+0000 mon.a (mon.0) 689 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.643321+0000 mon.a (mon.0) 690 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.643321+0000 mon.a (mon.0) 690 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: cephadm 2026-03-10T10:19:03.661636+0000 mgr.a (mgr.14150) 199 : cephadm [INF] Reconfiguring mgr.a (monmap changed)... 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: cephadm 2026-03-10T10:19:03.661636+0000 mgr.a (mgr.14150) 199 : cephadm [INF] Reconfiguring mgr.a (monmap changed)... 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.664208+0000 mon.a (mon.0) 691 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.664208+0000 mon.a (mon.0) 691 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.665056+0000 mon.a (mon.0) 692 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.665056+0000 mon.a (mon.0) 692 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.665594+0000 mon.a (mon.0) 693 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.665594+0000 mon.a (mon.0) 693 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: cephadm 2026-03-10T10:19:03.666536+0000 mgr.a (mgr.14150) 200 : cephadm [INF] Reconfiguring daemon mgr.a on vm00 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: cephadm 2026-03-10T10:19:03.666536+0000 mgr.a (mgr.14150) 200 : cephadm [INF] Reconfiguring daemon mgr.a on vm00 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.669658+0000 mon.a (mon.0) 694 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.669658+0000 mon.a (mon.0) 694 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.680779+0000 mon.a (mon.0) 695 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.680779+0000 mon.a (mon.0) 695 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.686238+0000 mon.a (mon.0) 696 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.686238+0000 mon.a (mon.0) 696 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.695126+0000 mon.a (mon.0) 697 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:03.695126+0000 mon.a (mon.0) 697 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:04.228112+0000 mon.a (mon.0) 698 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:04.228112+0000 mon.a (mon.0) 698 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:04.233794+0000 mon.a (mon.0) 699 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:04.233794+0000 mon.a (mon.0) 699 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:04.235007+0000 mon.a (mon.0) 700 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:04.235007+0000 mon.a (mon.0) 700 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:04.235469+0000 mon.a (mon.0) 701 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:05.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:04 vm03 bash[21274]: audit 2026-03-10T10:19:04.235469+0000 mon.a (mon.0) 701 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: cluster 2026-03-10T10:19:03.789490+0000 mgr.a (mgr.14150) 201 : cluster [DBG] pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: cluster 2026-03-10T10:19:03.789490+0000 mgr.a (mgr.14150) 201 : cluster [DBG] pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: cephadm 2026-03-10T10:19:04.234496+0000 mgr.a (mgr.14150) 202 : cephadm [INF] Reconfiguring osd.0 (monmap changed)... 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: cephadm 2026-03-10T10:19:04.234496+0000 mgr.a (mgr.14150) 202 : cephadm [INF] Reconfiguring osd.0 (monmap changed)... 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: cephadm 2026-03-10T10:19:04.236148+0000 mgr.a (mgr.14150) 203 : cephadm [INF] Reconfiguring daemon osd.0 on vm00 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: cephadm 2026-03-10T10:19:04.236148+0000 mgr.a (mgr.14150) 203 : cephadm [INF] Reconfiguring daemon osd.0 on vm00 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:04.704111+0000 mon.a (mon.0) 702 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:04.704111+0000 mon.a (mon.0) 702 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:04.709194+0000 mon.a (mon.0) 703 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:04.709194+0000 mon.a (mon.0) 703 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: cephadm 2026-03-10T10:19:04.709784+0000 mgr.a (mgr.14150) 204 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: cephadm 2026-03-10T10:19:04.709784+0000 mgr.a (mgr.14150) 204 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:04.710024+0000 mon.a (mon.0) 704 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:04.710024+0000 mon.a (mon.0) 704 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:04.710577+0000 mon.a (mon.0) 705 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:04.710577+0000 mon.a (mon.0) 705 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:04.711045+0000 mon.a (mon.0) 706 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:04.711045+0000 mon.a (mon.0) 706 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: cephadm 2026-03-10T10:19:04.711557+0000 mgr.a (mgr.14150) 205 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: cephadm 2026-03-10T10:19:04.711557+0000 mgr.a (mgr.14150) 205 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.169775+0000 mon.a (mon.0) 707 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.169775+0000 mon.a (mon.0) 707 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.173892+0000 mon.a (mon.0) 708 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.173892+0000 mon.a (mon.0) 708 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.174961+0000 mon.a (mon.0) 709 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.174961+0000 mon.a (mon.0) 709 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.175542+0000 mon.a (mon.0) 710 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.175542+0000 mon.a (mon.0) 710 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.176062+0000 mon.a (mon.0) 711 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.176062+0000 mon.a (mon.0) 711 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.582995+0000 mon.a (mon.0) 712 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.582995+0000 mon.a (mon.0) 712 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.587690+0000 mon.a (mon.0) 713 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.587690+0000 mon.a (mon.0) 713 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.589127+0000 mon.a (mon.0) 714 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.589127+0000 mon.a (mon.0) 714 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T10:19:06.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.589719+0000 mon.a (mon.0) 715 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.083 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:05 vm03 bash[21274]: audit 2026-03-10T10:19:05.589719+0000 mon.a (mon.0) 715 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: cluster 2026-03-10T10:19:03.789490+0000 mgr.a (mgr.14150) 201 : cluster [DBG] pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: cluster 2026-03-10T10:19:03.789490+0000 mgr.a (mgr.14150) 201 : cluster [DBG] pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: cephadm 2026-03-10T10:19:04.234496+0000 mgr.a (mgr.14150) 202 : cephadm [INF] Reconfiguring osd.0 (monmap changed)... 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: cephadm 2026-03-10T10:19:04.234496+0000 mgr.a (mgr.14150) 202 : cephadm [INF] Reconfiguring osd.0 (monmap changed)... 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: cephadm 2026-03-10T10:19:04.236148+0000 mgr.a (mgr.14150) 203 : cephadm [INF] Reconfiguring daemon osd.0 on vm00 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: cephadm 2026-03-10T10:19:04.236148+0000 mgr.a (mgr.14150) 203 : cephadm [INF] Reconfiguring daemon osd.0 on vm00 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:04.704111+0000 mon.a (mon.0) 702 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:04.704111+0000 mon.a (mon.0) 702 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:04.709194+0000 mon.a (mon.0) 703 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:04.709194+0000 mon.a (mon.0) 703 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: cephadm 2026-03-10T10:19:04.709784+0000 mgr.a (mgr.14150) 204 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: cephadm 2026-03-10T10:19:04.709784+0000 mgr.a (mgr.14150) 204 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:04.710024+0000 mon.a (mon.0) 704 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:04.710024+0000 mon.a (mon.0) 704 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:04.710577+0000 mon.a (mon.0) 705 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:04.710577+0000 mon.a (mon.0) 705 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:04.711045+0000 mon.a (mon.0) 706 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:04.711045+0000 mon.a (mon.0) 706 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: cephadm 2026-03-10T10:19:04.711557+0000 mgr.a (mgr.14150) 205 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:19:06.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: cephadm 2026-03-10T10:19:04.711557+0000 mgr.a (mgr.14150) 205 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.169775+0000 mon.a (mon.0) 707 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.169775+0000 mon.a (mon.0) 707 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.173892+0000 mon.a (mon.0) 708 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.173892+0000 mon.a (mon.0) 708 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.174961+0000 mon.a (mon.0) 709 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.174961+0000 mon.a (mon.0) 709 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.175542+0000 mon.a (mon.0) 710 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.175542+0000 mon.a (mon.0) 710 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.176062+0000 mon.a (mon.0) 711 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.176062+0000 mon.a (mon.0) 711 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.582995+0000 mon.a (mon.0) 712 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.582995+0000 mon.a (mon.0) 712 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.587690+0000 mon.a (mon.0) 713 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.587690+0000 mon.a (mon.0) 713 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.589127+0000 mon.a (mon.0) 714 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.589127+0000 mon.a (mon.0) 714 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.589719+0000 mon.a (mon.0) 715 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:05 vm00 bash[20709]: audit 2026-03-10T10:19:05.589719+0000 mon.a (mon.0) 715 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: cluster 2026-03-10T10:19:03.789490+0000 mgr.a (mgr.14150) 201 : cluster [DBG] pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: cluster 2026-03-10T10:19:03.789490+0000 mgr.a (mgr.14150) 201 : cluster [DBG] pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: cephadm 2026-03-10T10:19:04.234496+0000 mgr.a (mgr.14150) 202 : cephadm [INF] Reconfiguring osd.0 (monmap changed)... 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: cephadm 2026-03-10T10:19:04.234496+0000 mgr.a (mgr.14150) 202 : cephadm [INF] Reconfiguring osd.0 (monmap changed)... 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: cephadm 2026-03-10T10:19:04.236148+0000 mgr.a (mgr.14150) 203 : cephadm [INF] Reconfiguring daemon osd.0 on vm00 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: cephadm 2026-03-10T10:19:04.236148+0000 mgr.a (mgr.14150) 203 : cephadm [INF] Reconfiguring daemon osd.0 on vm00 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:04.704111+0000 mon.a (mon.0) 702 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:04.704111+0000 mon.a (mon.0) 702 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:04.709194+0000 mon.a (mon.0) 703 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:04.709194+0000 mon.a (mon.0) 703 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: cephadm 2026-03-10T10:19:04.709784+0000 mgr.a (mgr.14150) 204 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: cephadm 2026-03-10T10:19:04.709784+0000 mgr.a (mgr.14150) 204 : cephadm [INF] Reconfiguring mon.a (monmap changed)... 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:04.710024+0000 mon.a (mon.0) 704 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:04.710024+0000 mon.a (mon.0) 704 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:04.710577+0000 mon.a (mon.0) 705 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:04.710577+0000 mon.a (mon.0) 705 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:04.711045+0000 mon.a (mon.0) 706 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:04.711045+0000 mon.a (mon.0) 706 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: cephadm 2026-03-10T10:19:04.711557+0000 mgr.a (mgr.14150) 205 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: cephadm 2026-03-10T10:19:04.711557+0000 mgr.a (mgr.14150) 205 : cephadm [INF] Reconfiguring daemon mon.a on vm00 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.169775+0000 mon.a (mon.0) 707 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.169775+0000 mon.a (mon.0) 707 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.173892+0000 mon.a (mon.0) 708 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.173892+0000 mon.a (mon.0) 708 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.174961+0000 mon.a (mon.0) 709 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.174961+0000 mon.a (mon.0) 709 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.175542+0000 mon.a (mon.0) 710 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.175542+0000 mon.a (mon.0) 710 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.176062+0000 mon.a (mon.0) 711 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.176062+0000 mon.a (mon.0) 711 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.260 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.582995+0000 mon.a (mon.0) 712 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.260 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.582995+0000 mon.a (mon.0) 712 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.260 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.587690+0000 mon.a (mon.0) 713 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.260 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.587690+0000 mon.a (mon.0) 713 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:06.260 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.589127+0000 mon.a (mon.0) 714 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T10:19:06.260 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.589127+0000 mon.a (mon.0) 714 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T10:19:06.260 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.589719+0000 mon.a (mon.0) 715 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.260 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:05 vm09 bash[20916]: audit 2026-03-10T10:19:05.589719+0000 mon.a (mon.0) 715 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:06.770 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cephadm 2026-03-10T10:19:05.174692+0000 mgr.a (mgr.14150) 206 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cephadm 2026-03-10T10:19:05.174692+0000 mgr.a (mgr.14150) 206 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cephadm 2026-03-10T10:19:05.174692+0000 mgr.a (mgr.14150) 206 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cephadm 2026-03-10T10:19:05.176696+0000 mgr.a (mgr.14150) 207 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cephadm 2026-03-10T10:19:05.176696+0000 mgr.a (mgr.14150) 207 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cephadm 2026-03-10T10:19:05.588829+0000 mgr.a (mgr.14150) 208 : cephadm [INF] Reconfiguring osd.1 (monmap changed)... 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cephadm 2026-03-10T10:19:05.588829+0000 mgr.a (mgr.14150) 208 : cephadm [INF] Reconfiguring osd.1 (monmap changed)... 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cephadm 2026-03-10T10:19:05.590515+0000 mgr.a (mgr.14150) 209 : cephadm [INF] Reconfiguring daemon osd.1 on vm03 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cephadm 2026-03-10T10:19:05.590515+0000 mgr.a (mgr.14150) 209 : cephadm [INF] Reconfiguring daemon osd.1 on vm03 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cluster 2026-03-10T10:19:05.789745+0000 mgr.a (mgr.14150) 210 : cluster [DBG] pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cluster 2026-03-10T10:19:05.789745+0000 mgr.a (mgr.14150) 210 : cluster [DBG] pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.073391+0000 mon.a (mon.0) 716 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.073391+0000 mon.a (mon.0) 716 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.078078+0000 mon.a (mon.0) 717 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.078078+0000 mon.a (mon.0) 717 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cephadm 2026-03-10T10:19:06.078703+0000 mgr.a (mgr.14150) 211 : cephadm [INF] Reconfiguring mgr.b (monmap changed)... 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cephadm 2026-03-10T10:19:06.078703+0000 mgr.a (mgr.14150) 211 : cephadm [INF] Reconfiguring mgr.b (monmap changed)... 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.078895+0000 mon.a (mon.0) 718 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.078895+0000 mon.a (mon.0) 718 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.079400+0000 mon.a (mon.0) 719 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.079400+0000 mon.a (mon.0) 719 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:19:07.020 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.079758+0000 mon.a (mon.0) 720 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:07.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.079758+0000 mon.a (mon.0) 720 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:07.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cephadm 2026-03-10T10:19:06.080228+0000 mgr.a (mgr.14150) 212 : cephadm [INF] Reconfiguring daemon mgr.b on vm03 2026-03-10T10:19:07.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cephadm 2026-03-10T10:19:06.080228+0000 mgr.a (mgr.14150) 212 : cephadm [INF] Reconfiguring daemon mgr.b on vm03 2026-03-10T10:19:07.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.503753+0000 mon.a (mon.0) 721 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.503753+0000 mon.a (mon.0) 721 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.507901+0000 mon.a (mon.0) 722 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.507901+0000 mon.a (mon.0) 722 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cephadm 2026-03-10T10:19:06.508352+0000 mgr.a (mgr.14150) 213 : cephadm [INF] Reconfiguring osd.2 (monmap changed)... 2026-03-10T10:19:07.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cephadm 2026-03-10T10:19:06.508352+0000 mgr.a (mgr.14150) 213 : cephadm [INF] Reconfiguring osd.2 (monmap changed)... 2026-03-10T10:19:07.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.508507+0000 mon.a (mon.0) 723 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T10:19:07.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.508507+0000 mon.a (mon.0) 723 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T10:19:07.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.508934+0000 mon.a (mon.0) 724 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:07.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: audit 2026-03-10T10:19:06.508934+0000 mon.a (mon.0) 724 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:07.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cephadm 2026-03-10T10:19:06.509570+0000 mgr.a (mgr.14150) 214 : cephadm [INF] Reconfiguring daemon osd.2 on vm09 2026-03-10T10:19:07.021 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:06 vm09 bash[20916]: cephadm 2026-03-10T10:19:06.509570+0000 mgr.a (mgr.14150) 214 : cephadm [INF] Reconfiguring daemon osd.2 on vm09 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cephadm 2026-03-10T10:19:05.174692+0000 mgr.a (mgr.14150) 206 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cephadm 2026-03-10T10:19:05.176696+0000 mgr.a (mgr.14150) 207 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cephadm 2026-03-10T10:19:05.176696+0000 mgr.a (mgr.14150) 207 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cephadm 2026-03-10T10:19:05.588829+0000 mgr.a (mgr.14150) 208 : cephadm [INF] Reconfiguring osd.1 (monmap changed)... 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cephadm 2026-03-10T10:19:05.588829+0000 mgr.a (mgr.14150) 208 : cephadm [INF] Reconfiguring osd.1 (monmap changed)... 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cephadm 2026-03-10T10:19:05.590515+0000 mgr.a (mgr.14150) 209 : cephadm [INF] Reconfiguring daemon osd.1 on vm03 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cephadm 2026-03-10T10:19:05.590515+0000 mgr.a (mgr.14150) 209 : cephadm [INF] Reconfiguring daemon osd.1 on vm03 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cluster 2026-03-10T10:19:05.789745+0000 mgr.a (mgr.14150) 210 : cluster [DBG] pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cluster 2026-03-10T10:19:05.789745+0000 mgr.a (mgr.14150) 210 : cluster [DBG] pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.073391+0000 mon.a (mon.0) 716 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.073391+0000 mon.a (mon.0) 716 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.078078+0000 mon.a (mon.0) 717 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.078078+0000 mon.a (mon.0) 717 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cephadm 2026-03-10T10:19:06.078703+0000 mgr.a (mgr.14150) 211 : cephadm [INF] Reconfiguring mgr.b (monmap changed)... 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cephadm 2026-03-10T10:19:06.078703+0000 mgr.a (mgr.14150) 211 : cephadm [INF] Reconfiguring mgr.b (monmap changed)... 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.078895+0000 mon.a (mon.0) 718 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.078895+0000 mon.a (mon.0) 718 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.079400+0000 mon.a (mon.0) 719 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.079400+0000 mon.a (mon.0) 719 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.079758+0000 mon.a (mon.0) 720 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.079758+0000 mon.a (mon.0) 720 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cephadm 2026-03-10T10:19:06.080228+0000 mgr.a (mgr.14150) 212 : cephadm [INF] Reconfiguring daemon mgr.b on vm03 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cephadm 2026-03-10T10:19:06.080228+0000 mgr.a (mgr.14150) 212 : cephadm [INF] Reconfiguring daemon mgr.b on vm03 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.503753+0000 mon.a (mon.0) 721 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.503753+0000 mon.a (mon.0) 721 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.507901+0000 mon.a (mon.0) 722 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.507901+0000 mon.a (mon.0) 722 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cephadm 2026-03-10T10:19:06.508352+0000 mgr.a (mgr.14150) 213 : cephadm [INF] Reconfiguring osd.2 (monmap changed)... 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cephadm 2026-03-10T10:19:06.508352+0000 mgr.a (mgr.14150) 213 : cephadm [INF] Reconfiguring osd.2 (monmap changed)... 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.508507+0000 mon.a (mon.0) 723 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.508507+0000 mon.a (mon.0) 723 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.508934+0000 mon.a (mon.0) 724 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: audit 2026-03-10T10:19:06.508934+0000 mon.a (mon.0) 724 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cephadm 2026-03-10T10:19:06.509570+0000 mgr.a (mgr.14150) 214 : cephadm [INF] Reconfiguring daemon osd.2 on vm09 2026-03-10T10:19:07.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:06 vm03 bash[21274]: cephadm 2026-03-10T10:19:06.509570+0000 mgr.a (mgr.14150) 214 : cephadm [INF] Reconfiguring daemon osd.2 on vm09 2026-03-10T10:19:07.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cephadm 2026-03-10T10:19:05.174692+0000 mgr.a (mgr.14150) 206 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cephadm 2026-03-10T10:19:05.174692+0000 mgr.a (mgr.14150) 206 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cephadm 2026-03-10T10:19:05.176696+0000 mgr.a (mgr.14150) 207 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cephadm 2026-03-10T10:19:05.176696+0000 mgr.a (mgr.14150) 207 : cephadm [INF] Reconfiguring daemon mon.b on vm03 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cephadm 2026-03-10T10:19:05.588829+0000 mgr.a (mgr.14150) 208 : cephadm [INF] Reconfiguring osd.1 (monmap changed)... 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cephadm 2026-03-10T10:19:05.588829+0000 mgr.a (mgr.14150) 208 : cephadm [INF] Reconfiguring osd.1 (monmap changed)... 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cephadm 2026-03-10T10:19:05.590515+0000 mgr.a (mgr.14150) 209 : cephadm [INF] Reconfiguring daemon osd.1 on vm03 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cephadm 2026-03-10T10:19:05.590515+0000 mgr.a (mgr.14150) 209 : cephadm [INF] Reconfiguring daemon osd.1 on vm03 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cluster 2026-03-10T10:19:05.789745+0000 mgr.a (mgr.14150) 210 : cluster [DBG] pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cluster 2026-03-10T10:19:05.789745+0000 mgr.a (mgr.14150) 210 : cluster [DBG] pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.073391+0000 mon.a (mon.0) 716 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.073391+0000 mon.a (mon.0) 716 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.078078+0000 mon.a (mon.0) 717 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.078078+0000 mon.a (mon.0) 717 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cephadm 2026-03-10T10:19:06.078703+0000 mgr.a (mgr.14150) 211 : cephadm [INF] Reconfiguring mgr.b (monmap changed)... 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cephadm 2026-03-10T10:19:06.078703+0000 mgr.a (mgr.14150) 211 : cephadm [INF] Reconfiguring mgr.b (monmap changed)... 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.078895+0000 mon.a (mon.0) 718 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.078895+0000 mon.a (mon.0) 718 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.079400+0000 mon.a (mon.0) 719 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.079400+0000 mon.a (mon.0) 719 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.079758+0000 mon.a (mon.0) 720 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.079758+0000 mon.a (mon.0) 720 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cephadm 2026-03-10T10:19:06.080228+0000 mgr.a (mgr.14150) 212 : cephadm [INF] Reconfiguring daemon mgr.b on vm03 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cephadm 2026-03-10T10:19:06.080228+0000 mgr.a (mgr.14150) 212 : cephadm [INF] Reconfiguring daemon mgr.b on vm03 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.503753+0000 mon.a (mon.0) 721 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.503753+0000 mon.a (mon.0) 721 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.507901+0000 mon.a (mon.0) 722 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.507901+0000 mon.a (mon.0) 722 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cephadm 2026-03-10T10:19:06.508352+0000 mgr.a (mgr.14150) 213 : cephadm [INF] Reconfiguring osd.2 (monmap changed)... 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cephadm 2026-03-10T10:19:06.508352+0000 mgr.a (mgr.14150) 213 : cephadm [INF] Reconfiguring osd.2 (monmap changed)... 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.508507+0000 mon.a (mon.0) 723 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.508507+0000 mon.a (mon.0) 723 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.508934+0000 mon.a (mon.0) 724 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: audit 2026-03-10T10:19:06.508934+0000 mon.a (mon.0) 724 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cephadm 2026-03-10T10:19:06.509570+0000 mgr.a (mgr.14150) 214 : cephadm [INF] Reconfiguring daemon osd.2 on vm09 2026-03-10T10:19:07.167 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:06 vm00 bash[20709]: cephadm 2026-03-10T10:19:06.509570+0000 mgr.a (mgr.14150) 214 : cephadm [INF] Reconfiguring daemon osd.2 on vm09 2026-03-10T10:19:08.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:08 vm09 bash[20916]: audit 2026-03-10T10:19:07.428082+0000 mon.a (mon.0) 725 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:08.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:08 vm09 bash[20916]: audit 2026-03-10T10:19:07.428082+0000 mon.a (mon.0) 725 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:08.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:08 vm09 bash[20916]: audit 2026-03-10T10:19:07.507283+0000 mon.a (mon.0) 726 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:08.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:08 vm09 bash[20916]: audit 2026-03-10T10:19:07.507283+0000 mon.a (mon.0) 726 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:08.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:08 vm09 bash[20916]: cephadm 2026-03-10T10:19:07.508394+0000 mgr.a (mgr.14150) 215 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-10T10:19:08.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:08 vm09 bash[20916]: cephadm 2026-03-10T10:19:07.508394+0000 mgr.a (mgr.14150) 215 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-10T10:19:08.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:08 vm09 bash[20916]: audit 2026-03-10T10:19:07.508638+0000 mon.a (mon.0) 727 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:08.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:08 vm09 bash[20916]: audit 2026-03-10T10:19:07.508638+0000 mon.a (mon.0) 727 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:08.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:08 vm09 bash[20916]: audit 2026-03-10T10:19:07.509056+0000 mon.a (mon.0) 728 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:08.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:08 vm09 bash[20916]: audit 2026-03-10T10:19:07.509056+0000 mon.a (mon.0) 728 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:08.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:08 vm09 bash[20916]: audit 2026-03-10T10:19:07.509411+0000 mon.a (mon.0) 729 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:08.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:08 vm09 bash[20916]: audit 2026-03-10T10:19:07.509411+0000 mon.a (mon.0) 729 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:08.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:08 vm09 bash[20916]: cephadm 2026-03-10T10:19:07.509863+0000 mgr.a (mgr.14150) 216 : cephadm [INF] Reconfiguring daemon mon.c on vm09 2026-03-10T10:19:08.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:08 vm09 bash[20916]: cephadm 2026-03-10T10:19:07.509863+0000 mgr.a (mgr.14150) 216 : cephadm [INF] Reconfiguring daemon mon.c on vm09 2026-03-10T10:19:08.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:08 vm03 bash[21274]: audit 2026-03-10T10:19:07.428082+0000 mon.a (mon.0) 725 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:08.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:08 vm03 bash[21274]: audit 2026-03-10T10:19:07.428082+0000 mon.a (mon.0) 725 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:08.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:08 vm03 bash[21274]: audit 2026-03-10T10:19:07.507283+0000 mon.a (mon.0) 726 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:08.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:08 vm03 bash[21274]: audit 2026-03-10T10:19:07.507283+0000 mon.a (mon.0) 726 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:08.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:08 vm03 bash[21274]: cephadm 2026-03-10T10:19:07.508394+0000 mgr.a (mgr.14150) 215 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-10T10:19:08.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:08 vm03 bash[21274]: cephadm 2026-03-10T10:19:07.508394+0000 mgr.a (mgr.14150) 215 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-10T10:19:08.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:08 vm03 bash[21274]: audit 2026-03-10T10:19:07.508638+0000 mon.a (mon.0) 727 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:08.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:08 vm03 bash[21274]: audit 2026-03-10T10:19:07.508638+0000 mon.a (mon.0) 727 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:08.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:08 vm03 bash[21274]: audit 2026-03-10T10:19:07.509056+0000 mon.a (mon.0) 728 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:08.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:08 vm03 bash[21274]: audit 2026-03-10T10:19:07.509056+0000 mon.a (mon.0) 728 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:08.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:08 vm03 bash[21274]: audit 2026-03-10T10:19:07.509411+0000 mon.a (mon.0) 729 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:08.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:08 vm03 bash[21274]: audit 2026-03-10T10:19:07.509411+0000 mon.a (mon.0) 729 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:08.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:08 vm03 bash[21274]: cephadm 2026-03-10T10:19:07.509863+0000 mgr.a (mgr.14150) 216 : cephadm [INF] Reconfiguring daemon mon.c on vm09 2026-03-10T10:19:08.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:08 vm03 bash[21274]: cephadm 2026-03-10T10:19:07.509863+0000 mgr.a (mgr.14150) 216 : cephadm [INF] Reconfiguring daemon mon.c on vm09 2026-03-10T10:19:08.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:08 vm00 bash[20709]: audit 2026-03-10T10:19:07.428082+0000 mon.a (mon.0) 725 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:08.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:08 vm00 bash[20709]: audit 2026-03-10T10:19:07.428082+0000 mon.a (mon.0) 725 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:08.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:08 vm00 bash[20709]: audit 2026-03-10T10:19:07.507283+0000 mon.a (mon.0) 726 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:08.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:08 vm00 bash[20709]: audit 2026-03-10T10:19:07.507283+0000 mon.a (mon.0) 726 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:08.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:08 vm00 bash[20709]: cephadm 2026-03-10T10:19:07.508394+0000 mgr.a (mgr.14150) 215 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-10T10:19:08.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:08 vm00 bash[20709]: cephadm 2026-03-10T10:19:07.508394+0000 mgr.a (mgr.14150) 215 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-10T10:19:08.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:08 vm00 bash[20709]: audit 2026-03-10T10:19:07.508638+0000 mon.a (mon.0) 727 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:08.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:08 vm00 bash[20709]: audit 2026-03-10T10:19:07.508638+0000 mon.a (mon.0) 727 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:19:08.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:08 vm00 bash[20709]: audit 2026-03-10T10:19:07.509056+0000 mon.a (mon.0) 728 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:08.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:08 vm00 bash[20709]: audit 2026-03-10T10:19:07.509056+0000 mon.a (mon.0) 728 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:19:08.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:08 vm00 bash[20709]: audit 2026-03-10T10:19:07.509411+0000 mon.a (mon.0) 729 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:08.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:08 vm00 bash[20709]: audit 2026-03-10T10:19:07.509411+0000 mon.a (mon.0) 729 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:19:08.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:08 vm00 bash[20709]: cephadm 2026-03-10T10:19:07.509863+0000 mgr.a (mgr.14150) 216 : cephadm [INF] Reconfiguring daemon mon.c on vm09 2026-03-10T10:19:08.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:08 vm00 bash[20709]: cephadm 2026-03-10T10:19:07.509863+0000 mgr.a (mgr.14150) 216 : cephadm [INF] Reconfiguring daemon mon.c on vm09 2026-03-10T10:19:09.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:09 vm03 bash[21274]: cluster 2026-03-10T10:19:07.790004+0000 mgr.a (mgr.14150) 217 : cluster [DBG] pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:09.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:09 vm03 bash[21274]: cluster 2026-03-10T10:19:07.790004+0000 mgr.a (mgr.14150) 217 : cluster [DBG] pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:09.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:09 vm03 bash[21274]: audit 2026-03-10T10:19:08.465832+0000 mon.a (mon.0) 730 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:09.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:09 vm03 bash[21274]: audit 2026-03-10T10:19:08.465832+0000 mon.a (mon.0) 730 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:09.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:09 vm03 bash[21274]: audit 2026-03-10T10:19:08.470909+0000 mon.a (mon.0) 731 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:09.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:09 vm03 bash[21274]: audit 2026-03-10T10:19:08.470909+0000 mon.a (mon.0) 731 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:09.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:09 vm00 bash[20709]: cluster 2026-03-10T10:19:07.790004+0000 mgr.a (mgr.14150) 217 : cluster [DBG] pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:09.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:09 vm00 bash[20709]: cluster 2026-03-10T10:19:07.790004+0000 mgr.a (mgr.14150) 217 : cluster [DBG] pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:09.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:09 vm00 bash[20709]: audit 2026-03-10T10:19:08.465832+0000 mon.a (mon.0) 730 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:09.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:09 vm00 bash[20709]: audit 2026-03-10T10:19:08.465832+0000 mon.a (mon.0) 730 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:09.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:09 vm00 bash[20709]: audit 2026-03-10T10:19:08.470909+0000 mon.a (mon.0) 731 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:09.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:09 vm00 bash[20709]: audit 2026-03-10T10:19:08.470909+0000 mon.a (mon.0) 731 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:10.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:09 vm09 bash[20916]: cluster 2026-03-10T10:19:07.790004+0000 mgr.a (mgr.14150) 217 : cluster [DBG] pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:10.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:09 vm09 bash[20916]: cluster 2026-03-10T10:19:07.790004+0000 mgr.a (mgr.14150) 217 : cluster [DBG] pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:10.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:09 vm09 bash[20916]: audit 2026-03-10T10:19:08.465832+0000 mon.a (mon.0) 730 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:10.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:09 vm09 bash[20916]: audit 2026-03-10T10:19:08.465832+0000 mon.a (mon.0) 730 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:10.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:09 vm09 bash[20916]: audit 2026-03-10T10:19:08.470909+0000 mon.a (mon.0) 731 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:10.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:09 vm09 bash[20916]: audit 2026-03-10T10:19:08.470909+0000 mon.a (mon.0) 731 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:11.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:11 vm00 bash[20709]: cluster 2026-03-10T10:19:09.790311+0000 mgr.a (mgr.14150) 218 : cluster [DBG] pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:11.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:11 vm00 bash[20709]: cluster 2026-03-10T10:19:09.790311+0000 mgr.a (mgr.14150) 218 : cluster [DBG] pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:12.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:11 vm09 bash[20916]: cluster 2026-03-10T10:19:09.790311+0000 mgr.a (mgr.14150) 218 : cluster [DBG] pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:12.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:11 vm09 bash[20916]: cluster 2026-03-10T10:19:09.790311+0000 mgr.a (mgr.14150) 218 : cluster [DBG] pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:12.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:11 vm03 bash[21274]: cluster 2026-03-10T10:19:09.790311+0000 mgr.a (mgr.14150) 218 : cluster [DBG] pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:12.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:11 vm03 bash[21274]: cluster 2026-03-10T10:19:09.790311+0000 mgr.a (mgr.14150) 218 : cluster [DBG] pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:14.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:13 vm09 bash[20916]: cluster 2026-03-10T10:19:11.790574+0000 mgr.a (mgr.14150) 219 : cluster [DBG] pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:14.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:13 vm09 bash[20916]: cluster 2026-03-10T10:19:11.790574+0000 mgr.a (mgr.14150) 219 : cluster [DBG] pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:14.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:13 vm03 bash[21274]: cluster 2026-03-10T10:19:11.790574+0000 mgr.a (mgr.14150) 219 : cluster [DBG] pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:14.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:13 vm03 bash[21274]: cluster 2026-03-10T10:19:11.790574+0000 mgr.a (mgr.14150) 219 : cluster [DBG] pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:14.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:13 vm00 bash[20709]: cluster 2026-03-10T10:19:11.790574+0000 mgr.a (mgr.14150) 219 : cluster [DBG] pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:14.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:13 vm00 bash[20709]: cluster 2026-03-10T10:19:11.790574+0000 mgr.a (mgr.14150) 219 : cluster [DBG] pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:15.993 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:15 vm03 bash[21274]: cluster 2026-03-10T10:19:13.790885+0000 mgr.a (mgr.14150) 220 : cluster [DBG] pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:15.993 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:15 vm03 bash[21274]: cluster 2026-03-10T10:19:13.790885+0000 mgr.a (mgr.14150) 220 : cluster [DBG] pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:16.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:15 vm09 bash[20916]: cluster 2026-03-10T10:19:13.790885+0000 mgr.a (mgr.14150) 220 : cluster [DBG] pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:16.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:15 vm09 bash[20916]: cluster 2026-03-10T10:19:13.790885+0000 mgr.a (mgr.14150) 220 : cluster [DBG] pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:16.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:15 vm00 bash[20709]: cluster 2026-03-10T10:19:13.790885+0000 mgr.a (mgr.14150) 220 : cluster [DBG] pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:16.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:15 vm00 bash[20709]: cluster 2026-03-10T10:19:13.790885+0000 mgr.a (mgr.14150) 220 : cluster [DBG] pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:17.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:16 vm03 bash[21274]: cluster 2026-03-10T10:19:15.791121+0000 mgr.a (mgr.14150) 221 : cluster [DBG] pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:17.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:16 vm03 bash[21274]: cluster 2026-03-10T10:19:15.791121+0000 mgr.a (mgr.14150) 221 : cluster [DBG] pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:17.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:16 vm00 bash[20709]: cluster 2026-03-10T10:19:15.791121+0000 mgr.a (mgr.14150) 221 : cluster [DBG] pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:17.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:16 vm00 bash[20709]: cluster 2026-03-10T10:19:15.791121+0000 mgr.a (mgr.14150) 221 : cluster [DBG] pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:17.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:16 vm09 bash[20916]: cluster 2026-03-10T10:19:15.791121+0000 mgr.a (mgr.14150) 221 : cluster [DBG] pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:17.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:16 vm09 bash[20916]: cluster 2026-03-10T10:19:15.791121+0000 mgr.a (mgr.14150) 221 : cluster [DBG] pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:19.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:19 vm03 bash[21274]: cluster 2026-03-10T10:19:17.791425+0000 mgr.a (mgr.14150) 222 : cluster [DBG] pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:19.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:19 vm03 bash[21274]: cluster 2026-03-10T10:19:17.791425+0000 mgr.a (mgr.14150) 222 : cluster [DBG] pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:19.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:19 vm00 bash[20709]: cluster 2026-03-10T10:19:17.791425+0000 mgr.a (mgr.14150) 222 : cluster [DBG] pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:19.447 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:19 vm00 bash[20709]: cluster 2026-03-10T10:19:17.791425+0000 mgr.a (mgr.14150) 222 : cluster [DBG] pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:19.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:19 vm09 bash[20916]: cluster 2026-03-10T10:19:17.791425+0000 mgr.a (mgr.14150) 222 : cluster [DBG] pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:19.561 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:19 vm09 bash[20916]: cluster 2026-03-10T10:19:17.791425+0000 mgr.a (mgr.14150) 222 : cluster [DBG] pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:21.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:21 vm09 bash[20916]: cluster 2026-03-10T10:19:19.791750+0000 mgr.a (mgr.14150) 223 : cluster [DBG] pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:21.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:21 vm09 bash[20916]: cluster 2026-03-10T10:19:19.791750+0000 mgr.a (mgr.14150) 223 : cluster [DBG] pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:21.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:21 vm03 bash[21274]: cluster 2026-03-10T10:19:19.791750+0000 mgr.a (mgr.14150) 223 : cluster [DBG] pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:21.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:21 vm03 bash[21274]: cluster 2026-03-10T10:19:19.791750+0000 mgr.a (mgr.14150) 223 : cluster [DBG] pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:21.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:21 vm00 bash[20709]: cluster 2026-03-10T10:19:19.791750+0000 mgr.a (mgr.14150) 223 : cluster [DBG] pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:21.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:21 vm00 bash[20709]: cluster 2026-03-10T10:19:19.791750+0000 mgr.a (mgr.14150) 223 : cluster [DBG] pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:23.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:22 vm00 bash[20709]: audit 2026-03-10T10:19:21.744324+0000 mon.a (mon.0) 732 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:22 vm00 bash[20709]: audit 2026-03-10T10:19:21.744324+0000 mon.a (mon.0) 732 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:22 vm00 bash[20709]: audit 2026-03-10T10:19:21.754658+0000 mon.a (mon.0) 733 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:22 vm00 bash[20709]: audit 2026-03-10T10:19:21.754658+0000 mon.a (mon.0) 733 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:22 vm00 bash[20709]: audit 2026-03-10T10:19:21.759517+0000 mon.a (mon.0) 734 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:22 vm00 bash[20709]: audit 2026-03-10T10:19:21.759517+0000 mon.a (mon.0) 734 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:22 vm00 bash[20709]: cluster 2026-03-10T10:19:21.792010+0000 mgr.a (mgr.14150) 224 : cluster [DBG] pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:23.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:22 vm00 bash[20709]: cluster 2026-03-10T10:19:21.792010+0000 mgr.a (mgr.14150) 224 : cluster [DBG] pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:23.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:22 vm09 bash[20916]: audit 2026-03-10T10:19:21.744324+0000 mon.a (mon.0) 732 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:22 vm09 bash[20916]: audit 2026-03-10T10:19:21.744324+0000 mon.a (mon.0) 732 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:22 vm09 bash[20916]: audit 2026-03-10T10:19:21.754658+0000 mon.a (mon.0) 733 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:22 vm09 bash[20916]: audit 2026-03-10T10:19:21.754658+0000 mon.a (mon.0) 733 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:22 vm09 bash[20916]: audit 2026-03-10T10:19:21.759517+0000 mon.a (mon.0) 734 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:22 vm09 bash[20916]: audit 2026-03-10T10:19:21.759517+0000 mon.a (mon.0) 734 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:22 vm09 bash[20916]: cluster 2026-03-10T10:19:21.792010+0000 mgr.a (mgr.14150) 224 : cluster [DBG] pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:23.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:22 vm09 bash[20916]: cluster 2026-03-10T10:19:21.792010+0000 mgr.a (mgr.14150) 224 : cluster [DBG] pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:23.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:22 vm03 bash[21274]: audit 2026-03-10T10:19:21.744324+0000 mon.a (mon.0) 732 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:22 vm03 bash[21274]: audit 2026-03-10T10:19:21.744324+0000 mon.a (mon.0) 732 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:22 vm03 bash[21274]: audit 2026-03-10T10:19:21.754658+0000 mon.a (mon.0) 733 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:22 vm03 bash[21274]: audit 2026-03-10T10:19:21.754658+0000 mon.a (mon.0) 733 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:22 vm03 bash[21274]: audit 2026-03-10T10:19:21.759517+0000 mon.a (mon.0) 734 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:22 vm03 bash[21274]: audit 2026-03-10T10:19:21.759517+0000 mon.a (mon.0) 734 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:23.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:22 vm03 bash[21274]: cluster 2026-03-10T10:19:21.792010+0000 mgr.a (mgr.14150) 224 : cluster [DBG] pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:23.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:22 vm03 bash[21274]: cluster 2026-03-10T10:19:21.792010+0000 mgr.a (mgr.14150) 224 : cluster [DBG] pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:24.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:24 vm09 bash[20916]: audit 2026-03-10T10:19:22.812552+0000 mon.a (mon.0) 735 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:24 vm09 bash[20916]: audit 2026-03-10T10:19:22.812552+0000 mon.a (mon.0) 735 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:24 vm09 bash[20916]: audit 2026-03-10T10:19:22.979525+0000 mon.a (mon.0) 736 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:24 vm09 bash[20916]: audit 2026-03-10T10:19:22.979525+0000 mon.a (mon.0) 736 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:24 vm09 bash[20916]: audit 2026-03-10T10:19:23.034945+0000 mon.a (mon.0) 737 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:24 vm09 bash[20916]: audit 2026-03-10T10:19:23.034945+0000 mon.a (mon.0) 737 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:24 vm09 bash[20916]: audit 2026-03-10T10:19:23.849880+0000 mon.a (mon.0) 738 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:24 vm09 bash[20916]: audit 2026-03-10T10:19:23.849880+0000 mon.a (mon.0) 738 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:24 vm09 bash[20916]: audit 2026-03-10T10:19:23.880259+0000 mon.a (mon.0) 739 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:24 vm09 bash[20916]: audit 2026-03-10T10:19:23.880259+0000 mon.a (mon.0) 739 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:24 vm00 bash[20709]: audit 2026-03-10T10:19:22.812552+0000 mon.a (mon.0) 735 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:24 vm00 bash[20709]: audit 2026-03-10T10:19:22.812552+0000 mon.a (mon.0) 735 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:24 vm00 bash[20709]: audit 2026-03-10T10:19:22.979525+0000 mon.a (mon.0) 736 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:24 vm00 bash[20709]: audit 2026-03-10T10:19:22.979525+0000 mon.a (mon.0) 736 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:24 vm00 bash[20709]: audit 2026-03-10T10:19:23.034945+0000 mon.a (mon.0) 737 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:24 vm00 bash[20709]: audit 2026-03-10T10:19:23.034945+0000 mon.a (mon.0) 737 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:24 vm00 bash[20709]: audit 2026-03-10T10:19:23.849880+0000 mon.a (mon.0) 738 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:24 vm00 bash[20709]: audit 2026-03-10T10:19:23.849880+0000 mon.a (mon.0) 738 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:24 vm00 bash[20709]: audit 2026-03-10T10:19:23.880259+0000 mon.a (mon.0) 739 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.416 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:24 vm00 bash[20709]: audit 2026-03-10T10:19:23.880259+0000 mon.a (mon.0) 739 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:24 vm03 bash[21274]: audit 2026-03-10T10:19:22.812552+0000 mon.a (mon.0) 735 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:24 vm03 bash[21274]: audit 2026-03-10T10:19:22.812552+0000 mon.a (mon.0) 735 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:24 vm03 bash[21274]: audit 2026-03-10T10:19:22.979525+0000 mon.a (mon.0) 736 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:24 vm03 bash[21274]: audit 2026-03-10T10:19:22.979525+0000 mon.a (mon.0) 736 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:24 vm03 bash[21274]: audit 2026-03-10T10:19:23.034945+0000 mon.a (mon.0) 737 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:24 vm03 bash[21274]: audit 2026-03-10T10:19:23.034945+0000 mon.a (mon.0) 737 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:24 vm03 bash[21274]: audit 2026-03-10T10:19:23.849880+0000 mon.a (mon.0) 738 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:24 vm03 bash[21274]: audit 2026-03-10T10:19:23.849880+0000 mon.a (mon.0) 738 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:24 vm03 bash[21274]: audit 2026-03-10T10:19:23.880259+0000 mon.a (mon.0) 739 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:24.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:24 vm03 bash[21274]: audit 2026-03-10T10:19:23.880259+0000 mon.a (mon.0) 739 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:25.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:25 vm09 bash[20916]: cluster 2026-03-10T10:19:23.792212+0000 mgr.a (mgr.14150) 225 : cluster [DBG] pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:25.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:25 vm09 bash[20916]: cluster 2026-03-10T10:19:23.792212+0000 mgr.a (mgr.14150) 225 : cluster [DBG] pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:25.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:25 vm09 bash[20916]: audit 2026-03-10T10:19:24.002061+0000 mon.a (mon.0) 740 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:25.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:25 vm09 bash[20916]: audit 2026-03-10T10:19:24.002061+0000 mon.a (mon.0) 740 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:25.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:25 vm03 bash[21274]: cluster 2026-03-10T10:19:23.792212+0000 mgr.a (mgr.14150) 225 : cluster [DBG] pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:25.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:25 vm03 bash[21274]: cluster 2026-03-10T10:19:23.792212+0000 mgr.a (mgr.14150) 225 : cluster [DBG] pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:25.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:25 vm03 bash[21274]: audit 2026-03-10T10:19:24.002061+0000 mon.a (mon.0) 740 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:25.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:25 vm03 bash[21274]: audit 2026-03-10T10:19:24.002061+0000 mon.a (mon.0) 740 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:25.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:25 vm00 bash[20709]: cluster 2026-03-10T10:19:23.792212+0000 mgr.a (mgr.14150) 225 : cluster [DBG] pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:25.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:25 vm00 bash[20709]: cluster 2026-03-10T10:19:23.792212+0000 mgr.a (mgr.14150) 225 : cluster [DBG] pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:25.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:25 vm00 bash[20709]: audit 2026-03-10T10:19:24.002061+0000 mon.a (mon.0) 740 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:25.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:25 vm00 bash[20709]: audit 2026-03-10T10:19:24.002061+0000 mon.a (mon.0) 740 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:27.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:27 vm00 bash[20709]: cluster 2026-03-10T10:19:25.792473+0000 mgr.a (mgr.14150) 226 : cluster [DBG] pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:27.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:27 vm00 bash[20709]: cluster 2026-03-10T10:19:25.792473+0000 mgr.a (mgr.14150) 226 : cluster [DBG] pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:27.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:27 vm09 bash[20916]: cluster 2026-03-10T10:19:25.792473+0000 mgr.a (mgr.14150) 226 : cluster [DBG] pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:27.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:27 vm09 bash[20916]: cluster 2026-03-10T10:19:25.792473+0000 mgr.a (mgr.14150) 226 : cluster [DBG] pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:27.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:27 vm03 bash[21274]: cluster 2026-03-10T10:19:25.792473+0000 mgr.a (mgr.14150) 226 : cluster [DBG] pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:27.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:27 vm03 bash[21274]: cluster 2026-03-10T10:19:25.792473+0000 mgr.a (mgr.14150) 226 : cluster [DBG] pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:29.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:29 vm00 bash[20709]: cluster 2026-03-10T10:19:27.792856+0000 mgr.a (mgr.14150) 227 : cluster [DBG] pgmap v143: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:29.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:29 vm00 bash[20709]: cluster 2026-03-10T10:19:27.792856+0000 mgr.a (mgr.14150) 227 : cluster [DBG] pgmap v143: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:29.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:29 vm09 bash[20916]: cluster 2026-03-10T10:19:27.792856+0000 mgr.a (mgr.14150) 227 : cluster [DBG] pgmap v143: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:29.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:29 vm09 bash[20916]: cluster 2026-03-10T10:19:27.792856+0000 mgr.a (mgr.14150) 227 : cluster [DBG] pgmap v143: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:29.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:29 vm03 bash[21274]: cluster 2026-03-10T10:19:27.792856+0000 mgr.a (mgr.14150) 227 : cluster [DBG] pgmap v143: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:29.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:29 vm03 bash[21274]: cluster 2026-03-10T10:19:27.792856+0000 mgr.a (mgr.14150) 227 : cluster [DBG] pgmap v143: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:31.875 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:31 vm00 bash[20709]: cluster 2026-03-10T10:19:29.793220+0000 mgr.a (mgr.14150) 228 : cluster [DBG] pgmap v144: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:31.875 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:31 vm00 bash[20709]: cluster 2026-03-10T10:19:29.793220+0000 mgr.a (mgr.14150) 228 : cluster [DBG] pgmap v144: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:32.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:31 vm09 bash[20916]: cluster 2026-03-10T10:19:29.793220+0000 mgr.a (mgr.14150) 228 : cluster [DBG] pgmap v144: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:32.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:31 vm09 bash[20916]: cluster 2026-03-10T10:19:29.793220+0000 mgr.a (mgr.14150) 228 : cluster [DBG] pgmap v144: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:32.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:31 vm03 bash[21274]: cluster 2026-03-10T10:19:29.793220+0000 mgr.a (mgr.14150) 228 : cluster [DBG] pgmap v144: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:32.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:31 vm03 bash[21274]: cluster 2026-03-10T10:19:29.793220+0000 mgr.a (mgr.14150) 228 : cluster [DBG] pgmap v144: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:33.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:33 vm00 bash[20709]: cluster 2026-03-10T10:19:31.793533+0000 mgr.a (mgr.14150) 229 : cluster [DBG] pgmap v145: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:33.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:33 vm00 bash[20709]: cluster 2026-03-10T10:19:31.793533+0000 mgr.a (mgr.14150) 229 : cluster [DBG] pgmap v145: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:34.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:33 vm09 bash[20916]: cluster 2026-03-10T10:19:31.793533+0000 mgr.a (mgr.14150) 229 : cluster [DBG] pgmap v145: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:34.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:33 vm09 bash[20916]: cluster 2026-03-10T10:19:31.793533+0000 mgr.a (mgr.14150) 229 : cluster [DBG] pgmap v145: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:34.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:33 vm03 bash[21274]: cluster 2026-03-10T10:19:31.793533+0000 mgr.a (mgr.14150) 229 : cluster [DBG] pgmap v145: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:34.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:33 vm03 bash[21274]: cluster 2026-03-10T10:19:31.793533+0000 mgr.a (mgr.14150) 229 : cluster [DBG] pgmap v145: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:36.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:35 vm09 bash[20916]: cluster 2026-03-10T10:19:33.793969+0000 mgr.a (mgr.14150) 230 : cluster [DBG] pgmap v146: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:36.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:35 vm09 bash[20916]: cluster 2026-03-10T10:19:33.793969+0000 mgr.a (mgr.14150) 230 : cluster [DBG] pgmap v146: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:36.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:35 vm03 bash[21274]: cluster 2026-03-10T10:19:33.793969+0000 mgr.a (mgr.14150) 230 : cluster [DBG] pgmap v146: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:36.099 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:35 vm03 bash[21274]: cluster 2026-03-10T10:19:33.793969+0000 mgr.a (mgr.14150) 230 : cluster [DBG] pgmap v146: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:36.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:35 vm00 bash[20709]: cluster 2026-03-10T10:19:33.793969+0000 mgr.a (mgr.14150) 230 : cluster [DBG] pgmap v146: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:36.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:35 vm00 bash[20709]: cluster 2026-03-10T10:19:33.793969+0000 mgr.a (mgr.14150) 230 : cluster [DBG] pgmap v146: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:37.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:37 vm03 bash[21274]: cluster 2026-03-10T10:19:35.794521+0000 mgr.a (mgr.14150) 231 : cluster [DBG] pgmap v147: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:37.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:37 vm03 bash[21274]: cluster 2026-03-10T10:19:35.794521+0000 mgr.a (mgr.14150) 231 : cluster [DBG] pgmap v147: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:38.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:37 vm00 bash[20709]: cluster 2026-03-10T10:19:35.794521+0000 mgr.a (mgr.14150) 231 : cluster [DBG] pgmap v147: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:38.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:37 vm00 bash[20709]: cluster 2026-03-10T10:19:35.794521+0000 mgr.a (mgr.14150) 231 : cluster [DBG] pgmap v147: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:38.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:37 vm09 bash[20916]: cluster 2026-03-10T10:19:35.794521+0000 mgr.a (mgr.14150) 231 : cluster [DBG] pgmap v147: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:38.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:37 vm09 bash[20916]: cluster 2026-03-10T10:19:35.794521+0000 mgr.a (mgr.14150) 231 : cluster [DBG] pgmap v147: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:38.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:38 vm03 bash[21274]: cluster 2026-03-10T10:19:37.794822+0000 mgr.a (mgr.14150) 232 : cluster [DBG] pgmap v148: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:38.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:38 vm03 bash[21274]: cluster 2026-03-10T10:19:37.794822+0000 mgr.a (mgr.14150) 232 : cluster [DBG] pgmap v148: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:39.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:38 vm00 bash[20709]: cluster 2026-03-10T10:19:37.794822+0000 mgr.a (mgr.14150) 232 : cluster [DBG] pgmap v148: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:39.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:38 vm00 bash[20709]: cluster 2026-03-10T10:19:37.794822+0000 mgr.a (mgr.14150) 232 : cluster [DBG] pgmap v148: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:39.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:38 vm09 bash[20916]: cluster 2026-03-10T10:19:37.794822+0000 mgr.a (mgr.14150) 232 : cluster [DBG] pgmap v148: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:39.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:38 vm09 bash[20916]: cluster 2026-03-10T10:19:37.794822+0000 mgr.a (mgr.14150) 232 : cluster [DBG] pgmap v148: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:41.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:40 vm00 bash[20709]: cluster 2026-03-10T10:19:39.795065+0000 mgr.a (mgr.14150) 233 : cluster [DBG] pgmap v149: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:41.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:40 vm00 bash[20709]: cluster 2026-03-10T10:19:39.795065+0000 mgr.a (mgr.14150) 233 : cluster [DBG] pgmap v149: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:41.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:40 vm00 bash[20709]: audit 2026-03-10T10:19:39.847059+0000 mon.a (mon.0) 741 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:41.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:40 vm00 bash[20709]: audit 2026-03-10T10:19:39.847059+0000 mon.a (mon.0) 741 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:41.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:40 vm00 bash[20709]: audit 2026-03-10T10:19:39.919805+0000 mon.a (mon.0) 742 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:41.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:40 vm00 bash[20709]: audit 2026-03-10T10:19:39.919805+0000 mon.a (mon.0) 742 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:41.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:40 vm00 bash[20709]: audit 2026-03-10T10:19:39.941927+0000 mon.a (mon.0) 743 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:41.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:40 vm00 bash[20709]: audit 2026-03-10T10:19:39.941927+0000 mon.a (mon.0) 743 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:41.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:40 vm09 bash[20916]: cluster 2026-03-10T10:19:39.795065+0000 mgr.a (mgr.14150) 233 : cluster [DBG] pgmap v149: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:41.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:40 vm09 bash[20916]: cluster 2026-03-10T10:19:39.795065+0000 mgr.a (mgr.14150) 233 : cluster [DBG] pgmap v149: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:41.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:40 vm09 bash[20916]: audit 2026-03-10T10:19:39.847059+0000 mon.a (mon.0) 741 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:41.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:40 vm09 bash[20916]: audit 2026-03-10T10:19:39.847059+0000 mon.a (mon.0) 741 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:41.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:40 vm09 bash[20916]: audit 2026-03-10T10:19:39.919805+0000 mon.a (mon.0) 742 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:41.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:40 vm09 bash[20916]: audit 2026-03-10T10:19:39.919805+0000 mon.a (mon.0) 742 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:41.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:40 vm09 bash[20916]: audit 2026-03-10T10:19:39.941927+0000 mon.a (mon.0) 743 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:41.259 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:40 vm09 bash[20916]: audit 2026-03-10T10:19:39.941927+0000 mon.a (mon.0) 743 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:41.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:40 vm03 bash[21274]: cluster 2026-03-10T10:19:39.795065+0000 mgr.a (mgr.14150) 233 : cluster [DBG] pgmap v149: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:41.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:40 vm03 bash[21274]: cluster 2026-03-10T10:19:39.795065+0000 mgr.a (mgr.14150) 233 : cluster [DBG] pgmap v149: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:41.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:40 vm03 bash[21274]: audit 2026-03-10T10:19:39.847059+0000 mon.a (mon.0) 741 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:41.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:40 vm03 bash[21274]: audit 2026-03-10T10:19:39.847059+0000 mon.a (mon.0) 741 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:41.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:40 vm03 bash[21274]: audit 2026-03-10T10:19:39.919805+0000 mon.a (mon.0) 742 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:41.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:40 vm03 bash[21274]: audit 2026-03-10T10:19:39.919805+0000 mon.a (mon.0) 742 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:41.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:40 vm03 bash[21274]: audit 2026-03-10T10:19:39.941927+0000 mon.a (mon.0) 743 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:41.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:40 vm03 bash[21274]: audit 2026-03-10T10:19:39.941927+0000 mon.a (mon.0) 743 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:42 vm09 bash[20916]: audit 2026-03-10T10:19:41.126122+0000 mon.a (mon.0) 744 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:42 vm09 bash[20916]: audit 2026-03-10T10:19:41.126122+0000 mon.a (mon.0) 744 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:42 vm09 bash[20916]: audit 2026-03-10T10:19:41.183877+0000 mon.a (mon.0) 745 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:42 vm09 bash[20916]: audit 2026-03-10T10:19:41.183877+0000 mon.a (mon.0) 745 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:42 vm09 bash[20916]: audit 2026-03-10T10:19:41.196179+0000 mon.a (mon.0) 746 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:42 vm09 bash[20916]: audit 2026-03-10T10:19:41.196179+0000 mon.a (mon.0) 746 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:42 vm03 bash[21274]: audit 2026-03-10T10:19:41.126122+0000 mon.a (mon.0) 744 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:42 vm03 bash[21274]: audit 2026-03-10T10:19:41.126122+0000 mon.a (mon.0) 744 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:42 vm03 bash[21274]: audit 2026-03-10T10:19:41.183877+0000 mon.a (mon.0) 745 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:42 vm03 bash[21274]: audit 2026-03-10T10:19:41.183877+0000 mon.a (mon.0) 745 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:42 vm03 bash[21274]: audit 2026-03-10T10:19:41.196179+0000 mon.a (mon.0) 746 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:42 vm03 bash[21274]: audit 2026-03-10T10:19:41.196179+0000 mon.a (mon.0) 746 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:42 vm00 bash[20709]: audit 2026-03-10T10:19:41.126122+0000 mon.a (mon.0) 744 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:42 vm00 bash[20709]: audit 2026-03-10T10:19:41.126122+0000 mon.a (mon.0) 744 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:42 vm00 bash[20709]: audit 2026-03-10T10:19:41.183877+0000 mon.a (mon.0) 745 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:42 vm00 bash[20709]: audit 2026-03-10T10:19:41.183877+0000 mon.a (mon.0) 745 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:42 vm00 bash[20709]: audit 2026-03-10T10:19:41.196179+0000 mon.a (mon.0) 746 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:42.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:42 vm00 bash[20709]: audit 2026-03-10T10:19:41.196179+0000 mon.a (mon.0) 746 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:43.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:43 vm03 bash[21274]: cluster 2026-03-10T10:19:41.795298+0000 mgr.a (mgr.14150) 234 : cluster [DBG] pgmap v150: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:43.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:43 vm03 bash[21274]: cluster 2026-03-10T10:19:41.795298+0000 mgr.a (mgr.14150) 234 : cluster [DBG] pgmap v150: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:43.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:43 vm00 bash[20709]: cluster 2026-03-10T10:19:41.795298+0000 mgr.a (mgr.14150) 234 : cluster [DBG] pgmap v150: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:43.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:43 vm00 bash[20709]: cluster 2026-03-10T10:19:41.795298+0000 mgr.a (mgr.14150) 234 : cluster [DBG] pgmap v150: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:43.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:43 vm09 bash[20916]: cluster 2026-03-10T10:19:41.795298+0000 mgr.a (mgr.14150) 234 : cluster [DBG] pgmap v150: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:43.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:43 vm09 bash[20916]: cluster 2026-03-10T10:19:41.795298+0000 mgr.a (mgr.14150) 234 : cluster [DBG] pgmap v150: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:45.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:45 vm09 bash[20916]: cluster 2026-03-10T10:19:43.795676+0000 mgr.a (mgr.14150) 235 : cluster [DBG] pgmap v151: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:45.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:45 vm09 bash[20916]: cluster 2026-03-10T10:19:43.795676+0000 mgr.a (mgr.14150) 235 : cluster [DBG] pgmap v151: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:45.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:45 vm09 bash[20916]: audit 2026-03-10T10:19:44.144954+0000 mon.a (mon.0) 747 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:45.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:45 vm09 bash[20916]: audit 2026-03-10T10:19:44.144954+0000 mon.a (mon.0) 747 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:45.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:45 vm09 bash[20916]: audit 2026-03-10T10:19:44.202679+0000 mon.a (mon.0) 748 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:45.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:45 vm09 bash[20916]: audit 2026-03-10T10:19:44.202679+0000 mon.a (mon.0) 748 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:45.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:45 vm09 bash[20916]: audit 2026-03-10T10:19:44.239125+0000 mon.a (mon.0) 749 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:45.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:45 vm09 bash[20916]: audit 2026-03-10T10:19:44.239125+0000 mon.a (mon.0) 749 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:45.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:45 vm03 bash[21274]: cluster 2026-03-10T10:19:43.795676+0000 mgr.a (mgr.14150) 235 : cluster [DBG] pgmap v151: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:45.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:45 vm03 bash[21274]: cluster 2026-03-10T10:19:43.795676+0000 mgr.a (mgr.14150) 235 : cluster [DBG] pgmap v151: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:45.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:45 vm03 bash[21274]: audit 2026-03-10T10:19:44.144954+0000 mon.a (mon.0) 747 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:45.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:45 vm03 bash[21274]: audit 2026-03-10T10:19:44.144954+0000 mon.a (mon.0) 747 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:45.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:45 vm03 bash[21274]: audit 2026-03-10T10:19:44.202679+0000 mon.a (mon.0) 748 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:45.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:45 vm03 bash[21274]: audit 2026-03-10T10:19:44.202679+0000 mon.a (mon.0) 748 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:45.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:45 vm03 bash[21274]: audit 2026-03-10T10:19:44.239125+0000 mon.a (mon.0) 749 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:45.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:45 vm03 bash[21274]: audit 2026-03-10T10:19:44.239125+0000 mon.a (mon.0) 749 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:45.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:45 vm00 bash[20709]: cluster 2026-03-10T10:19:43.795676+0000 mgr.a (mgr.14150) 235 : cluster [DBG] pgmap v151: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:45.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:45 vm00 bash[20709]: cluster 2026-03-10T10:19:43.795676+0000 mgr.a (mgr.14150) 235 : cluster [DBG] pgmap v151: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:45.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:45 vm00 bash[20709]: audit 2026-03-10T10:19:44.144954+0000 mon.a (mon.0) 747 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:45.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:45 vm00 bash[20709]: audit 2026-03-10T10:19:44.144954+0000 mon.a (mon.0) 747 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:45.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:45 vm00 bash[20709]: audit 2026-03-10T10:19:44.202679+0000 mon.a (mon.0) 748 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:45.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:45 vm00 bash[20709]: audit 2026-03-10T10:19:44.202679+0000 mon.a (mon.0) 748 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:45.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:45 vm00 bash[20709]: audit 2026-03-10T10:19:44.239125+0000 mon.a (mon.0) 749 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:45.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:45 vm00 bash[20709]: audit 2026-03-10T10:19:44.239125+0000 mon.a (mon.0) 749 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:47.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:47 vm09 bash[20916]: cluster 2026-03-10T10:19:45.796048+0000 mgr.a (mgr.14150) 236 : cluster [DBG] pgmap v152: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:47.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:47 vm09 bash[20916]: cluster 2026-03-10T10:19:45.796048+0000 mgr.a (mgr.14150) 236 : cluster [DBG] pgmap v152: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:47.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:47 vm03 bash[21274]: cluster 2026-03-10T10:19:45.796048+0000 mgr.a (mgr.14150) 236 : cluster [DBG] pgmap v152: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:47.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:47 vm03 bash[21274]: cluster 2026-03-10T10:19:45.796048+0000 mgr.a (mgr.14150) 236 : cluster [DBG] pgmap v152: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:47.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:47 vm00 bash[20709]: cluster 2026-03-10T10:19:45.796048+0000 mgr.a (mgr.14150) 236 : cluster [DBG] pgmap v152: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:47.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:47 vm00 bash[20709]: cluster 2026-03-10T10:19:45.796048+0000 mgr.a (mgr.14150) 236 : cluster [DBG] pgmap v152: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:49.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:49 vm09 bash[20916]: cluster 2026-03-10T10:19:47.796363+0000 mgr.a (mgr.14150) 237 : cluster [DBG] pgmap v153: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:49.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:49 vm09 bash[20916]: cluster 2026-03-10T10:19:47.796363+0000 mgr.a (mgr.14150) 237 : cluster [DBG] pgmap v153: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:49.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:49 vm03 bash[21274]: cluster 2026-03-10T10:19:47.796363+0000 mgr.a (mgr.14150) 237 : cluster [DBG] pgmap v153: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:49.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:49 vm03 bash[21274]: cluster 2026-03-10T10:19:47.796363+0000 mgr.a (mgr.14150) 237 : cluster [DBG] pgmap v153: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:49.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:49 vm00 bash[20709]: cluster 2026-03-10T10:19:47.796363+0000 mgr.a (mgr.14150) 237 : cluster [DBG] pgmap v153: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:49.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:49 vm00 bash[20709]: cluster 2026-03-10T10:19:47.796363+0000 mgr.a (mgr.14150) 237 : cluster [DBG] pgmap v153: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:51.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:51 vm09 bash[20916]: cluster 2026-03-10T10:19:49.796668+0000 mgr.a (mgr.14150) 238 : cluster [DBG] pgmap v154: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:51.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:51 vm09 bash[20916]: cluster 2026-03-10T10:19:49.796668+0000 mgr.a (mgr.14150) 238 : cluster [DBG] pgmap v154: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:51.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:51 vm03 bash[21274]: cluster 2026-03-10T10:19:49.796668+0000 mgr.a (mgr.14150) 238 : cluster [DBG] pgmap v154: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:51.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:51 vm03 bash[21274]: cluster 2026-03-10T10:19:49.796668+0000 mgr.a (mgr.14150) 238 : cluster [DBG] pgmap v154: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:51.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:51 vm00 bash[20709]: cluster 2026-03-10T10:19:49.796668+0000 mgr.a (mgr.14150) 238 : cluster [DBG] pgmap v154: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:51.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:51 vm00 bash[20709]: cluster 2026-03-10T10:19:49.796668+0000 mgr.a (mgr.14150) 238 : cluster [DBG] pgmap v154: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:53.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:53 vm09 bash[20916]: cluster 2026-03-10T10:19:51.796941+0000 mgr.a (mgr.14150) 239 : cluster [DBG] pgmap v155: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:53.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:53 vm09 bash[20916]: cluster 2026-03-10T10:19:51.796941+0000 mgr.a (mgr.14150) 239 : cluster [DBG] pgmap v155: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:53.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:53 vm03 bash[21274]: cluster 2026-03-10T10:19:51.796941+0000 mgr.a (mgr.14150) 239 : cluster [DBG] pgmap v155: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:53.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:53 vm03 bash[21274]: cluster 2026-03-10T10:19:51.796941+0000 mgr.a (mgr.14150) 239 : cluster [DBG] pgmap v155: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:53.652 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:53 vm00 bash[20709]: cluster 2026-03-10T10:19:51.796941+0000 mgr.a (mgr.14150) 239 : cluster [DBG] pgmap v155: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:53.652 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:53 vm00 bash[20709]: cluster 2026-03-10T10:19:51.796941+0000 mgr.a (mgr.14150) 239 : cluster [DBG] pgmap v155: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:56.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:55 vm09 bash[20916]: cluster 2026-03-10T10:19:53.797233+0000 mgr.a (mgr.14150) 240 : cluster [DBG] pgmap v156: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:56.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:55 vm09 bash[20916]: cluster 2026-03-10T10:19:53.797233+0000 mgr.a (mgr.14150) 240 : cluster [DBG] pgmap v156: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:56.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:55 vm03 bash[21274]: cluster 2026-03-10T10:19:53.797233+0000 mgr.a (mgr.14150) 240 : cluster [DBG] pgmap v156: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:56.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:55 vm03 bash[21274]: cluster 2026-03-10T10:19:53.797233+0000 mgr.a (mgr.14150) 240 : cluster [DBG] pgmap v156: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:56.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:55 vm00 bash[20709]: cluster 2026-03-10T10:19:53.797233+0000 mgr.a (mgr.14150) 240 : cluster [DBG] pgmap v156: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:56.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:55 vm00 bash[20709]: cluster 2026-03-10T10:19:53.797233+0000 mgr.a (mgr.14150) 240 : cluster [DBG] pgmap v156: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:58.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:57 vm09 bash[20916]: cluster 2026-03-10T10:19:55.797483+0000 mgr.a (mgr.14150) 241 : cluster [DBG] pgmap v157: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:58.009 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:57 vm09 bash[20916]: cluster 2026-03-10T10:19:55.797483+0000 mgr.a (mgr.14150) 241 : cluster [DBG] pgmap v157: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:58.036 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:57 vm03 bash[21274]: cluster 2026-03-10T10:19:55.797483+0000 mgr.a (mgr.14150) 241 : cluster [DBG] pgmap v157: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:58.036 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:57 vm03 bash[21274]: cluster 2026-03-10T10:19:55.797483+0000 mgr.a (mgr.14150) 241 : cluster [DBG] pgmap v157: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:58.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:57 vm00 bash[20709]: cluster 2026-03-10T10:19:55.797483+0000 mgr.a (mgr.14150) 241 : cluster [DBG] pgmap v157: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:58.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:57 vm00 bash[20709]: cluster 2026-03-10T10:19:55.797483+0000 mgr.a (mgr.14150) 241 : cluster [DBG] pgmap v157: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:59.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:59 vm00 bash[20709]: cluster 2026-03-10T10:19:57.797810+0000 mgr.a (mgr.14150) 242 : cluster [DBG] pgmap v158: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:59.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:59 vm00 bash[20709]: cluster 2026-03-10T10:19:57.797810+0000 mgr.a (mgr.14150) 242 : cluster [DBG] pgmap v158: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:59.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:59 vm00 bash[20709]: audit 2026-03-10T10:19:58.024114+0000 mon.a (mon.0) 750 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:59.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:59 vm00 bash[20709]: audit 2026-03-10T10:19:58.024114+0000 mon.a (mon.0) 750 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:59.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:59 vm00 bash[20709]: audit 2026-03-10T10:19:58.029643+0000 mon.a (mon.0) 751 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:59.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:59 vm00 bash[20709]: audit 2026-03-10T10:19:58.029643+0000 mon.a (mon.0) 751 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:59.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:59 vm00 bash[20709]: audit 2026-03-10T10:19:58.036195+0000 mon.a (mon.0) 752 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:59.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:19:59 vm00 bash[20709]: audit 2026-03-10T10:19:58.036195+0000 mon.a (mon.0) 752 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:59.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:59 vm03 bash[21274]: cluster 2026-03-10T10:19:57.797810+0000 mgr.a (mgr.14150) 242 : cluster [DBG] pgmap v158: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:59.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:59 vm03 bash[21274]: cluster 2026-03-10T10:19:57.797810+0000 mgr.a (mgr.14150) 242 : cluster [DBG] pgmap v158: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:59.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:59 vm03 bash[21274]: audit 2026-03-10T10:19:58.024114+0000 mon.a (mon.0) 750 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:59.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:59 vm03 bash[21274]: audit 2026-03-10T10:19:58.024114+0000 mon.a (mon.0) 750 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:59.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:59 vm03 bash[21274]: audit 2026-03-10T10:19:58.029643+0000 mon.a (mon.0) 751 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:59.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:59 vm03 bash[21274]: audit 2026-03-10T10:19:58.029643+0000 mon.a (mon.0) 751 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:59.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:59 vm03 bash[21274]: audit 2026-03-10T10:19:58.036195+0000 mon.a (mon.0) 752 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:59.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:19:59 vm03 bash[21274]: audit 2026-03-10T10:19:58.036195+0000 mon.a (mon.0) 752 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:59.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:59 vm09 bash[20916]: cluster 2026-03-10T10:19:57.797810+0000 mgr.a (mgr.14150) 242 : cluster [DBG] pgmap v158: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:59.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:59 vm09 bash[20916]: cluster 2026-03-10T10:19:57.797810+0000 mgr.a (mgr.14150) 242 : cluster [DBG] pgmap v158: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:19:59.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:59 vm09 bash[20916]: audit 2026-03-10T10:19:58.024114+0000 mon.a (mon.0) 750 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:59.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:59 vm09 bash[20916]: audit 2026-03-10T10:19:58.024114+0000 mon.a (mon.0) 750 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:59.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:59 vm09 bash[20916]: audit 2026-03-10T10:19:58.029643+0000 mon.a (mon.0) 751 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:59.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:59 vm09 bash[20916]: audit 2026-03-10T10:19:58.029643+0000 mon.a (mon.0) 751 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:59.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:59 vm09 bash[20916]: audit 2026-03-10T10:19:58.036195+0000 mon.a (mon.0) 752 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:19:59.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:19:59 vm09 bash[20916]: audit 2026-03-10T10:19:58.036195+0000 mon.a (mon.0) 752 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:00 vm03 bash[21274]: audit 2026-03-10T10:19:59.267900+0000 mon.a (mon.0) 753 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:00 vm03 bash[21274]: audit 2026-03-10T10:19:59.267900+0000 mon.a (mon.0) 753 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:00 vm03 bash[21274]: audit 2026-03-10T10:19:59.273620+0000 mon.a (mon.0) 754 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:00 vm03 bash[21274]: audit 2026-03-10T10:19:59.273620+0000 mon.a (mon.0) 754 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:00 vm03 bash[21274]: audit 2026-03-10T10:19:59.279249+0000 mon.a (mon.0) 755 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:00 vm03 bash[21274]: audit 2026-03-10T10:19:59.279249+0000 mon.a (mon.0) 755 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:00 vm03 bash[21274]: cluster 2026-03-10T10:20:00.000181+0000 mon.a (mon.0) 756 : cluster [INF] overall HEALTH_OK 2026-03-10T10:20:00.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:00 vm03 bash[21274]: cluster 2026-03-10T10:20:00.000181+0000 mon.a (mon.0) 756 : cluster [INF] overall HEALTH_OK 2026-03-10T10:20:00.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:00 vm00 bash[20709]: audit 2026-03-10T10:19:59.267900+0000 mon.a (mon.0) 753 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:00 vm00 bash[20709]: audit 2026-03-10T10:19:59.267900+0000 mon.a (mon.0) 753 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:00 vm00 bash[20709]: audit 2026-03-10T10:19:59.273620+0000 mon.a (mon.0) 754 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:00 vm00 bash[20709]: audit 2026-03-10T10:19:59.273620+0000 mon.a (mon.0) 754 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:00 vm00 bash[20709]: audit 2026-03-10T10:19:59.279249+0000 mon.a (mon.0) 755 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:00 vm00 bash[20709]: audit 2026-03-10T10:19:59.279249+0000 mon.a (mon.0) 755 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:00 vm00 bash[20709]: cluster 2026-03-10T10:20:00.000181+0000 mon.a (mon.0) 756 : cluster [INF] overall HEALTH_OK 2026-03-10T10:20:00.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:00 vm00 bash[20709]: cluster 2026-03-10T10:20:00.000181+0000 mon.a (mon.0) 756 : cluster [INF] overall HEALTH_OK 2026-03-10T10:20:00.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:00 vm09 bash[20916]: audit 2026-03-10T10:19:59.267900+0000 mon.a (mon.0) 753 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:00 vm09 bash[20916]: audit 2026-03-10T10:19:59.267900+0000 mon.a (mon.0) 753 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:00 vm09 bash[20916]: audit 2026-03-10T10:19:59.273620+0000 mon.a (mon.0) 754 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:00 vm09 bash[20916]: audit 2026-03-10T10:19:59.273620+0000 mon.a (mon.0) 754 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:00 vm09 bash[20916]: audit 2026-03-10T10:19:59.279249+0000 mon.a (mon.0) 755 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:00 vm09 bash[20916]: audit 2026-03-10T10:19:59.279249+0000 mon.a (mon.0) 755 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:00.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:00 vm09 bash[20916]: cluster 2026-03-10T10:20:00.000181+0000 mon.a (mon.0) 756 : cluster [INF] overall HEALTH_OK 2026-03-10T10:20:00.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:00 vm09 bash[20916]: cluster 2026-03-10T10:20:00.000181+0000 mon.a (mon.0) 756 : cluster [INF] overall HEALTH_OK 2026-03-10T10:20:01.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:01 vm03 bash[21274]: cluster 2026-03-10T10:19:59.798050+0000 mgr.a (mgr.14150) 243 : cluster [DBG] pgmap v159: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:01.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:01 vm03 bash[21274]: cluster 2026-03-10T10:19:59.798050+0000 mgr.a (mgr.14150) 243 : cluster [DBG] pgmap v159: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:01.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:01 vm00 bash[20709]: cluster 2026-03-10T10:19:59.798050+0000 mgr.a (mgr.14150) 243 : cluster [DBG] pgmap v159: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:01.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:01 vm00 bash[20709]: cluster 2026-03-10T10:19:59.798050+0000 mgr.a (mgr.14150) 243 : cluster [DBG] pgmap v159: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:01.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:01 vm09 bash[20916]: cluster 2026-03-10T10:19:59.798050+0000 mgr.a (mgr.14150) 243 : cluster [DBG] pgmap v159: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:01.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:01 vm09 bash[20916]: cluster 2026-03-10T10:19:59.798050+0000 mgr.a (mgr.14150) 243 : cluster [DBG] pgmap v159: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:03.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:03 vm09 bash[20916]: cluster 2026-03-10T10:20:01.798272+0000 mgr.a (mgr.14150) 244 : cluster [DBG] pgmap v160: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:03.509 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:03 vm09 bash[20916]: cluster 2026-03-10T10:20:01.798272+0000 mgr.a (mgr.14150) 244 : cluster [DBG] pgmap v160: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:03.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:03 vm03 bash[21274]: cluster 2026-03-10T10:20:01.798272+0000 mgr.a (mgr.14150) 244 : cluster [DBG] pgmap v160: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:03.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:03 vm03 bash[21274]: cluster 2026-03-10T10:20:01.798272+0000 mgr.a (mgr.14150) 244 : cluster [DBG] pgmap v160: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:03.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:03 vm00 bash[20709]: cluster 2026-03-10T10:20:01.798272+0000 mgr.a (mgr.14150) 244 : cluster [DBG] pgmap v160: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:03.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:03 vm00 bash[20709]: cluster 2026-03-10T10:20:01.798272+0000 mgr.a (mgr.14150) 244 : cluster [DBG] pgmap v160: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:05.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:05 vm00 bash[20709]: cluster 2026-03-10T10:20:03.798516+0000 mgr.a (mgr.14150) 245 : cluster [DBG] pgmap v161: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:05.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:05 vm00 bash[20709]: cluster 2026-03-10T10:20:03.798516+0000 mgr.a (mgr.14150) 245 : cluster [DBG] pgmap v161: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:05.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:05 vm00 bash[20709]: audit 2026-03-10T10:20:04.319205+0000 mon.a (mon.0) 757 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:05.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:05 vm00 bash[20709]: audit 2026-03-10T10:20:04.319205+0000 mon.a (mon.0) 757 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:05.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:05 vm00 bash[20709]: audit 2026-03-10T10:20:04.323185+0000 mon.a (mon.0) 758 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:05.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:05 vm00 bash[20709]: audit 2026-03-10T10:20:04.323185+0000 mon.a (mon.0) 758 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:05.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:05 vm00 bash[20709]: audit 2026-03-10T10:20:04.327185+0000 mon.a (mon.0) 759 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:05.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:05 vm00 bash[20709]: audit 2026-03-10T10:20:04.327185+0000 mon.a (mon.0) 759 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:05.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:05 vm09 bash[20916]: cluster 2026-03-10T10:20:03.798516+0000 mgr.a (mgr.14150) 245 : cluster [DBG] pgmap v161: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:05.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:05 vm09 bash[20916]: cluster 2026-03-10T10:20:03.798516+0000 mgr.a (mgr.14150) 245 : cluster [DBG] pgmap v161: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:05.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:05 vm09 bash[20916]: audit 2026-03-10T10:20:04.319205+0000 mon.a (mon.0) 757 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:05.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:05 vm09 bash[20916]: audit 2026-03-10T10:20:04.319205+0000 mon.a (mon.0) 757 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:05.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:05 vm09 bash[20916]: audit 2026-03-10T10:20:04.323185+0000 mon.a (mon.0) 758 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:05.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:05 vm09 bash[20916]: audit 2026-03-10T10:20:04.323185+0000 mon.a (mon.0) 758 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:05.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:05 vm09 bash[20916]: audit 2026-03-10T10:20:04.327185+0000 mon.a (mon.0) 759 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:05.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:05 vm09 bash[20916]: audit 2026-03-10T10:20:04.327185+0000 mon.a (mon.0) 759 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:05.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:05 vm03 bash[21274]: cluster 2026-03-10T10:20:03.798516+0000 mgr.a (mgr.14150) 245 : cluster [DBG] pgmap v161: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:05.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:05 vm03 bash[21274]: cluster 2026-03-10T10:20:03.798516+0000 mgr.a (mgr.14150) 245 : cluster [DBG] pgmap v161: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:05.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:05 vm03 bash[21274]: audit 2026-03-10T10:20:04.319205+0000 mon.a (mon.0) 757 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:05.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:05 vm03 bash[21274]: audit 2026-03-10T10:20:04.319205+0000 mon.a (mon.0) 757 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:05.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:05 vm03 bash[21274]: audit 2026-03-10T10:20:04.323185+0000 mon.a (mon.0) 758 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:05.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:05 vm03 bash[21274]: audit 2026-03-10T10:20:04.323185+0000 mon.a (mon.0) 758 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:05.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:05 vm03 bash[21274]: audit 2026-03-10T10:20:04.327185+0000 mon.a (mon.0) 759 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:05.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:05 vm03 bash[21274]: audit 2026-03-10T10:20:04.327185+0000 mon.a (mon.0) 759 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:07.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:07 vm00 bash[20709]: cluster 2026-03-10T10:20:05.798724+0000 mgr.a (mgr.14150) 246 : cluster [DBG] pgmap v162: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:07.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:07 vm00 bash[20709]: cluster 2026-03-10T10:20:05.798724+0000 mgr.a (mgr.14150) 246 : cluster [DBG] pgmap v162: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:07.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:07 vm09 bash[20916]: cluster 2026-03-10T10:20:05.798724+0000 mgr.a (mgr.14150) 246 : cluster [DBG] pgmap v162: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:07.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:07 vm09 bash[20916]: cluster 2026-03-10T10:20:05.798724+0000 mgr.a (mgr.14150) 246 : cluster [DBG] pgmap v162: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:07.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:07 vm03 bash[21274]: cluster 2026-03-10T10:20:05.798724+0000 mgr.a (mgr.14150) 246 : cluster [DBG] pgmap v162: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:07.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:07 vm03 bash[21274]: cluster 2026-03-10T10:20:05.798724+0000 mgr.a (mgr.14150) 246 : cluster [DBG] pgmap v162: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:09.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:09 vm09 bash[20916]: cluster 2026-03-10T10:20:07.799027+0000 mgr.a (mgr.14150) 247 : cluster [DBG] pgmap v163: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:09.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:09 vm09 bash[20916]: cluster 2026-03-10T10:20:07.799027+0000 mgr.a (mgr.14150) 247 : cluster [DBG] pgmap v163: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:09.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:09 vm09 bash[20916]: audit 2026-03-10T10:20:08.472015+0000 mon.a (mon.0) 760 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:09.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:09 vm09 bash[20916]: audit 2026-03-10T10:20:08.472015+0000 mon.a (mon.0) 760 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:09.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:09 vm09 bash[20916]: audit 2026-03-10T10:20:08.473622+0000 mon.a (mon.0) 761 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:09.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:09 vm09 bash[20916]: audit 2026-03-10T10:20:08.473622+0000 mon.a (mon.0) 761 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:09.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:09 vm09 bash[20916]: audit 2026-03-10T10:20:08.474148+0000 mon.a (mon.0) 762 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:09.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:09 vm09 bash[20916]: audit 2026-03-10T10:20:08.474148+0000 mon.a (mon.0) 762 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:09.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:09 vm09 bash[20916]: audit 2026-03-10T10:20:08.478909+0000 mon.a (mon.0) 763 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:09.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:09 vm09 bash[20916]: audit 2026-03-10T10:20:08.478909+0000 mon.a (mon.0) 763 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:09.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:09 vm03 bash[21274]: cluster 2026-03-10T10:20:07.799027+0000 mgr.a (mgr.14150) 247 : cluster [DBG] pgmap v163: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:09.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:09 vm03 bash[21274]: cluster 2026-03-10T10:20:07.799027+0000 mgr.a (mgr.14150) 247 : cluster [DBG] pgmap v163: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:09.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:09 vm03 bash[21274]: audit 2026-03-10T10:20:08.472015+0000 mon.a (mon.0) 760 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:09.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:09 vm03 bash[21274]: audit 2026-03-10T10:20:08.472015+0000 mon.a (mon.0) 760 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:09.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:09 vm03 bash[21274]: audit 2026-03-10T10:20:08.473622+0000 mon.a (mon.0) 761 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:09.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:09 vm03 bash[21274]: audit 2026-03-10T10:20:08.473622+0000 mon.a (mon.0) 761 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:09.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:09 vm03 bash[21274]: audit 2026-03-10T10:20:08.474148+0000 mon.a (mon.0) 762 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:09.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:09 vm03 bash[21274]: audit 2026-03-10T10:20:08.474148+0000 mon.a (mon.0) 762 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:09.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:09 vm03 bash[21274]: audit 2026-03-10T10:20:08.478909+0000 mon.a (mon.0) 763 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:09.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:09 vm03 bash[21274]: audit 2026-03-10T10:20:08.478909+0000 mon.a (mon.0) 763 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:09.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:09 vm00 bash[20709]: cluster 2026-03-10T10:20:07.799027+0000 mgr.a (mgr.14150) 247 : cluster [DBG] pgmap v163: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:09.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:09 vm00 bash[20709]: cluster 2026-03-10T10:20:07.799027+0000 mgr.a (mgr.14150) 247 : cluster [DBG] pgmap v163: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:09.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:09 vm00 bash[20709]: audit 2026-03-10T10:20:08.472015+0000 mon.a (mon.0) 760 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:09.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:09 vm00 bash[20709]: audit 2026-03-10T10:20:08.472015+0000 mon.a (mon.0) 760 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:09.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:09 vm00 bash[20709]: audit 2026-03-10T10:20:08.473622+0000 mon.a (mon.0) 761 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:09.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:09 vm00 bash[20709]: audit 2026-03-10T10:20:08.473622+0000 mon.a (mon.0) 761 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:09.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:09 vm00 bash[20709]: audit 2026-03-10T10:20:08.474148+0000 mon.a (mon.0) 762 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:09.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:09 vm00 bash[20709]: audit 2026-03-10T10:20:08.474148+0000 mon.a (mon.0) 762 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:09.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:09 vm00 bash[20709]: audit 2026-03-10T10:20:08.478909+0000 mon.a (mon.0) 763 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:09.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:09 vm00 bash[20709]: audit 2026-03-10T10:20:08.478909+0000 mon.a (mon.0) 763 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:11.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:11 vm09 bash[20916]: cluster 2026-03-10T10:20:09.799288+0000 mgr.a (mgr.14150) 248 : cluster [DBG] pgmap v164: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:11.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:11 vm09 bash[20916]: cluster 2026-03-10T10:20:09.799288+0000 mgr.a (mgr.14150) 248 : cluster [DBG] pgmap v164: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:11.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:11 vm03 bash[21274]: cluster 2026-03-10T10:20:09.799288+0000 mgr.a (mgr.14150) 248 : cluster [DBG] pgmap v164: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:11.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:11 vm03 bash[21274]: cluster 2026-03-10T10:20:09.799288+0000 mgr.a (mgr.14150) 248 : cluster [DBG] pgmap v164: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:11.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:11 vm00 bash[20709]: cluster 2026-03-10T10:20:09.799288+0000 mgr.a (mgr.14150) 248 : cluster [DBG] pgmap v164: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:11.920 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:11 vm00 bash[20709]: cluster 2026-03-10T10:20:09.799288+0000 mgr.a (mgr.14150) 248 : cluster [DBG] pgmap v164: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:13.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:13 vm00 bash[20709]: cluster 2026-03-10T10:20:11.799568+0000 mgr.a (mgr.14150) 249 : cluster [DBG] pgmap v165: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:13.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:13 vm00 bash[20709]: cluster 2026-03-10T10:20:11.799568+0000 mgr.a (mgr.14150) 249 : cluster [DBG] pgmap v165: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:13.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:13 vm09 bash[20916]: cluster 2026-03-10T10:20:11.799568+0000 mgr.a (mgr.14150) 249 : cluster [DBG] pgmap v165: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:13.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:13 vm09 bash[20916]: cluster 2026-03-10T10:20:11.799568+0000 mgr.a (mgr.14150) 249 : cluster [DBG] pgmap v165: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:13.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:13 vm03 bash[21274]: cluster 2026-03-10T10:20:11.799568+0000 mgr.a (mgr.14150) 249 : cluster [DBG] pgmap v165: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:13.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:13 vm03 bash[21274]: cluster 2026-03-10T10:20:11.799568+0000 mgr.a (mgr.14150) 249 : cluster [DBG] pgmap v165: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:15.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:15 vm09 bash[20916]: cluster 2026-03-10T10:20:13.799977+0000 mgr.a (mgr.14150) 250 : cluster [DBG] pgmap v166: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:15.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:15 vm09 bash[20916]: cluster 2026-03-10T10:20:13.799977+0000 mgr.a (mgr.14150) 250 : cluster [DBG] pgmap v166: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:15.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:15 vm03 bash[21274]: cluster 2026-03-10T10:20:13.799977+0000 mgr.a (mgr.14150) 250 : cluster [DBG] pgmap v166: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:15.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:15 vm03 bash[21274]: cluster 2026-03-10T10:20:13.799977+0000 mgr.a (mgr.14150) 250 : cluster [DBG] pgmap v166: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:15.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:15 vm00 bash[20709]: cluster 2026-03-10T10:20:13.799977+0000 mgr.a (mgr.14150) 250 : cluster [DBG] pgmap v166: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:15.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:15 vm00 bash[20709]: cluster 2026-03-10T10:20:13.799977+0000 mgr.a (mgr.14150) 250 : cluster [DBG] pgmap v166: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:17.493 INFO:teuthology.orchestra.run.vm00.stderr:+ ceph orch ps --refresh 2026-03-10T10:20:17.654 INFO:teuthology.orchestra.run.vm00.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-10T10:20:17.654 INFO:teuthology.orchestra.run.vm00.stdout:agent.vm00 vm00 running 18s ago 5m - - 2026-03-10T10:20:17.654 INFO:teuthology.orchestra.run.vm00.stdout:agent.vm03 vm03 running 19s ago 4m - - 2026-03-10T10:20:17.654 INFO:teuthology.orchestra.run.vm00.stdout:agent.vm09 vm09 running 13s ago 4m - - 2026-03-10T10:20:17.654 INFO:teuthology.orchestra.run.vm00.stdout:mgr.a vm00 *:9283,8765 running (5m) 18s ago 5m 527M - 19.2.3-678-ge911bdeb 654f31e6858e d78dd054fe4e 2026-03-10T10:20:17.654 INFO:teuthology.orchestra.run.vm00.stdout:mgr.b vm03 *:8443,8765 running (4m) 19s ago 4m 462M - 19.2.3-678-ge911bdeb 654f31e6858e 8b51996eb83f 2026-03-10T10:20:17.654 INFO:teuthology.orchestra.run.vm00.stdout:mon.a vm00 running (5m) 18s ago 5m 41.7M 2048M 19.2.3-678-ge911bdeb 654f31e6858e df7968fa808a 2026-03-10T10:20:17.654 INFO:teuthology.orchestra.run.vm00.stdout:mon.b vm03 running (4m) 19s ago 4m 40.3M 2048M 19.2.3-678-ge911bdeb 654f31e6858e bfc71159537c 2026-03-10T10:20:17.654 INFO:teuthology.orchestra.run.vm00.stdout:mon.c vm09 running (4m) 13s ago 4m 36.1M 2048M 19.2.3-678-ge911bdeb 654f31e6858e 7ada9fb041b4 2026-03-10T10:20:17.654 INFO:teuthology.orchestra.run.vm00.stdout:osd.0 vm00 running (3m) 18s ago 3m 35.7M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 7ab8d200666f 2026-03-10T10:20:17.654 INFO:teuthology.orchestra.run.vm00.stdout:osd.1 vm03 running (3m) 19s ago 3m 36.5M 4096M 19.2.3-678-ge911bdeb 654f31e6858e ddd2db8f86f8 2026-03-10T10:20:17.654 INFO:teuthology.orchestra.run.vm00.stdout:osd.2 vm09 running (2m) 13s ago 2m 35.4M 3527M 19.2.3-678-ge911bdeb 654f31e6858e 90385710fb52 2026-03-10T10:20:17.675 INFO:teuthology.orchestra.run.vm00.stderr:+ ceph orch ls --service-name mon --export 2026-03-10T10:20:17.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:17 vm09 bash[20916]: cluster 2026-03-10T10:20:15.800246+0000 mgr.a (mgr.14150) 251 : cluster [DBG] pgmap v167: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:17.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:17 vm09 bash[20916]: cluster 2026-03-10T10:20:15.800246+0000 mgr.a (mgr.14150) 251 : cluster [DBG] pgmap v167: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:17.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:17 vm03 bash[21274]: cluster 2026-03-10T10:20:15.800246+0000 mgr.a (mgr.14150) 251 : cluster [DBG] pgmap v167: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:17.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:17 vm03 bash[21274]: cluster 2026-03-10T10:20:15.800246+0000 mgr.a (mgr.14150) 251 : cluster [DBG] pgmap v167: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:17.844 INFO:teuthology.orchestra.run.vm00.stderr:+ ceph orch apply -i mon.yaml 2026-03-10T10:20:17.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:17 vm00 bash[20709]: cluster 2026-03-10T10:20:15.800246+0000 mgr.a (mgr.14150) 251 : cluster [DBG] pgmap v167: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:17.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:17 vm00 bash[20709]: cluster 2026-03-10T10:20:15.800246+0000 mgr.a (mgr.14150) 251 : cluster [DBG] pgmap v167: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:18.027 INFO:teuthology.orchestra.run.vm00.stdout:Scheduled mon update... 2026-03-10T10:20:18.051 INFO:teuthology.orchestra.run.vm00.stderr:+ sleep 90 2026-03-10T10:20:18.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:17.651796+0000 mgr.a (mgr.14150) 252 : audit [DBG] from='client.24274 -' entity='client.admin' cmd=[{"prefix": "orch ps", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:18.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:17.651796+0000 mgr.a (mgr.14150) 252 : audit [DBG] from='client.24274 -' entity='client.admin' cmd=[{"prefix": "orch ps", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:18.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:17.655106+0000 mon.a (mon.0) 764 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:17.655106+0000 mon.a (mon.0) 764 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:17.656867+0000 mon.a (mon.0) 765 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:17.656867+0000 mon.a (mon.0) 765 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:17.657326+0000 mon.a (mon.0) 766 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:17.657326+0000 mon.a (mon.0) 766 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:17.661984+0000 mon.a (mon.0) 767 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:17.661984+0000 mon.a (mon.0) 767 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.026461+0000 mon.a (mon.0) 768 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.026461+0000 mon.a (mon.0) 768 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.027687+0000 mon.a (mon.0) 769 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.027687+0000 mon.a (mon.0) 769 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.030018+0000 mon.a (mon.0) 770 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.030018+0000 mon.a (mon.0) 770 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.030944+0000 mon.a (mon.0) 771 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.030944+0000 mon.a (mon.0) 771 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.037635+0000 mon.a (mon.0) 772 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.037635+0000 mon.a (mon.0) 772 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.039689+0000 mon.a (mon.0) 773 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.039689+0000 mon.a (mon.0) 773 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.045896+0000 mon.a (mon.0) 774 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.045896+0000 mon.a (mon.0) 774 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.102512+0000 mon.a (mon.0) 775 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.102512+0000 mon.a (mon.0) 775 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.107559+0000 mon.a (mon.0) 776 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.107559+0000 mon.a (mon.0) 776 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.113493+0000 mon.a (mon.0) 777 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.759 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:18 vm09 bash[20916]: audit 2026-03-10T10:20:18.113493+0000 mon.a (mon.0) 777 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:17.651796+0000 mgr.a (mgr.14150) 252 : audit [DBG] from='client.24274 -' entity='client.admin' cmd=[{"prefix": "orch ps", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:17.651796+0000 mgr.a (mgr.14150) 252 : audit [DBG] from='client.24274 -' entity='client.admin' cmd=[{"prefix": "orch ps", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:17.655106+0000 mon.a (mon.0) 764 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:17.655106+0000 mon.a (mon.0) 764 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:17.656867+0000 mon.a (mon.0) 765 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:17.656867+0000 mon.a (mon.0) 765 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:17.657326+0000 mon.a (mon.0) 766 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:17.657326+0000 mon.a (mon.0) 766 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:17.661984+0000 mon.a (mon.0) 767 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:17.661984+0000 mon.a (mon.0) 767 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.026461+0000 mon.a (mon.0) 768 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.026461+0000 mon.a (mon.0) 768 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.027687+0000 mon.a (mon.0) 769 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.027687+0000 mon.a (mon.0) 769 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.030018+0000 mon.a (mon.0) 770 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.030018+0000 mon.a (mon.0) 770 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.030944+0000 mon.a (mon.0) 771 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.030944+0000 mon.a (mon.0) 771 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.037635+0000 mon.a (mon.0) 772 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.037635+0000 mon.a (mon.0) 772 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.039689+0000 mon.a (mon.0) 773 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.039689+0000 mon.a (mon.0) 773 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.045896+0000 mon.a (mon.0) 774 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.045896+0000 mon.a (mon.0) 774 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.102512+0000 mon.a (mon.0) 775 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.102512+0000 mon.a (mon.0) 775 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.107559+0000 mon.a (mon.0) 776 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.107559+0000 mon.a (mon.0) 776 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.113493+0000 mon.a (mon.0) 777 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.832 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:18 vm03 bash[21274]: audit 2026-03-10T10:20:18.113493+0000 mon.a (mon.0) 777 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:17.651796+0000 mgr.a (mgr.14150) 252 : audit [DBG] from='client.24274 -' entity='client.admin' cmd=[{"prefix": "orch ps", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:17.651796+0000 mgr.a (mgr.14150) 252 : audit [DBG] from='client.24274 -' entity='client.admin' cmd=[{"prefix": "orch ps", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:17.655106+0000 mon.a (mon.0) 764 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:17.655106+0000 mon.a (mon.0) 764 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:17.656867+0000 mon.a (mon.0) 765 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:17.656867+0000 mon.a (mon.0) 765 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:17.657326+0000 mon.a (mon.0) 766 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:17.657326+0000 mon.a (mon.0) 766 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:17.661984+0000 mon.a (mon.0) 767 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:17.661984+0000 mon.a (mon.0) 767 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.026461+0000 mon.a (mon.0) 768 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.026461+0000 mon.a (mon.0) 768 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.027687+0000 mon.a (mon.0) 769 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.027687+0000 mon.a (mon.0) 769 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.030018+0000 mon.a (mon.0) 770 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.030018+0000 mon.a (mon.0) 770 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.030944+0000 mon.a (mon.0) 771 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.030944+0000 mon.a (mon.0) 771 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.037635+0000 mon.a (mon.0) 772 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.037635+0000 mon.a (mon.0) 772 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.039689+0000 mon.a (mon.0) 773 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.039689+0000 mon.a (mon.0) 773 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "quorum_status"}]: dispatch 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.045896+0000 mon.a (mon.0) 774 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.045896+0000 mon.a (mon.0) 774 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.102512+0000 mon.a (mon.0) 775 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.102512+0000 mon.a (mon.0) 775 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.107559+0000 mon.a (mon.0) 776 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.107559+0000 mon.a (mon.0) 776 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.113493+0000 mon.a (mon.0) 777 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:18.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:18 vm00 bash[20709]: audit 2026-03-10T10:20:18.113493+0000 mon.a (mon.0) 777 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:19 vm09 bash[20916]: cluster 2026-03-10T10:20:17.800556+0000 mgr.a (mgr.14150) 253 : cluster [DBG] pgmap v168: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:20.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:19 vm09 bash[20916]: cluster 2026-03-10T10:20:17.800556+0000 mgr.a (mgr.14150) 253 : cluster [DBG] pgmap v168: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:20.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:19 vm09 bash[20916]: audit 2026-03-10T10:20:17.832024+0000 mgr.a (mgr.14150) 254 : audit [DBG] from='client.14409 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:20.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:19 vm09 bash[20916]: audit 2026-03-10T10:20:17.832024+0000 mgr.a (mgr.14150) 254 : audit [DBG] from='client.14409 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:20.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:19 vm09 bash[20916]: audit 2026-03-10T10:20:18.001380+0000 mgr.a (mgr.14150) 255 : audit [DBG] from='client.14415 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:20.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:19 vm09 bash[20916]: audit 2026-03-10T10:20:18.001380+0000 mgr.a (mgr.14150) 255 : audit [DBG] from='client.14415 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:20.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:19 vm09 bash[20916]: cephadm 2026-03-10T10:20:18.002986+0000 mgr.a (mgr.14150) 256 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:20:20.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:19 vm09 bash[20916]: cephadm 2026-03-10T10:20:18.002986+0000 mgr.a (mgr.14150) 256 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:20:20.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:19 vm09 bash[20916]: audit 2026-03-10T10:20:19.399297+0000 mon.a (mon.0) 778 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:19 vm09 bash[20916]: audit 2026-03-10T10:20:19.399297+0000 mon.a (mon.0) 778 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:19 vm09 bash[20916]: audit 2026-03-10T10:20:19.433933+0000 mon.a (mon.0) 779 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:19 vm09 bash[20916]: audit 2026-03-10T10:20:19.433933+0000 mon.a (mon.0) 779 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:19 vm09 bash[20916]: audit 2026-03-10T10:20:19.470731+0000 mon.a (mon.0) 780 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:19 vm09 bash[20916]: audit 2026-03-10T10:20:19.470731+0000 mon.a (mon.0) 780 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:19 vm03 bash[21274]: cluster 2026-03-10T10:20:17.800556+0000 mgr.a (mgr.14150) 253 : cluster [DBG] pgmap v168: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:20.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:19 vm03 bash[21274]: cluster 2026-03-10T10:20:17.800556+0000 mgr.a (mgr.14150) 253 : cluster [DBG] pgmap v168: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:20.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:19 vm03 bash[21274]: audit 2026-03-10T10:20:17.832024+0000 mgr.a (mgr.14150) 254 : audit [DBG] from='client.14409 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:20.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:19 vm03 bash[21274]: audit 2026-03-10T10:20:17.832024+0000 mgr.a (mgr.14150) 254 : audit [DBG] from='client.14409 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:20.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:19 vm03 bash[21274]: audit 2026-03-10T10:20:18.001380+0000 mgr.a (mgr.14150) 255 : audit [DBG] from='client.14415 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:20.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:19 vm03 bash[21274]: audit 2026-03-10T10:20:18.001380+0000 mgr.a (mgr.14150) 255 : audit [DBG] from='client.14415 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:20.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:19 vm03 bash[21274]: cephadm 2026-03-10T10:20:18.002986+0000 mgr.a (mgr.14150) 256 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:20:20.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:19 vm03 bash[21274]: cephadm 2026-03-10T10:20:18.002986+0000 mgr.a (mgr.14150) 256 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:20:20.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:19 vm03 bash[21274]: audit 2026-03-10T10:20:19.399297+0000 mon.a (mon.0) 778 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:19 vm03 bash[21274]: audit 2026-03-10T10:20:19.399297+0000 mon.a (mon.0) 778 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:19 vm03 bash[21274]: audit 2026-03-10T10:20:19.433933+0000 mon.a (mon.0) 779 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:19 vm03 bash[21274]: audit 2026-03-10T10:20:19.433933+0000 mon.a (mon.0) 779 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:19 vm03 bash[21274]: audit 2026-03-10T10:20:19.470731+0000 mon.a (mon.0) 780 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:19 vm03 bash[21274]: audit 2026-03-10T10:20:19.470731+0000 mon.a (mon.0) 780 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:19 vm00 bash[20709]: cluster 2026-03-10T10:20:17.800556+0000 mgr.a (mgr.14150) 253 : cluster [DBG] pgmap v168: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:20.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:19 vm00 bash[20709]: cluster 2026-03-10T10:20:17.800556+0000 mgr.a (mgr.14150) 253 : cluster [DBG] pgmap v168: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:20.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:19 vm00 bash[20709]: audit 2026-03-10T10:20:17.832024+0000 mgr.a (mgr.14150) 254 : audit [DBG] from='client.14409 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:20.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:19 vm00 bash[20709]: audit 2026-03-10T10:20:17.832024+0000 mgr.a (mgr.14150) 254 : audit [DBG] from='client.14409 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_name": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:20.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:19 vm00 bash[20709]: audit 2026-03-10T10:20:18.001380+0000 mgr.a (mgr.14150) 255 : audit [DBG] from='client.14415 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:20.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:19 vm00 bash[20709]: audit 2026-03-10T10:20:18.001380+0000 mgr.a (mgr.14150) 255 : audit [DBG] from='client.14415 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:20:20.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:19 vm00 bash[20709]: cephadm 2026-03-10T10:20:18.002986+0000 mgr.a (mgr.14150) 256 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:20:20.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:19 vm00 bash[20709]: cephadm 2026-03-10T10:20:18.002986+0000 mgr.a (mgr.14150) 256 : cephadm [INF] Saving service mon spec with placement count:3 2026-03-10T10:20:20.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:19 vm00 bash[20709]: audit 2026-03-10T10:20:19.399297+0000 mon.a (mon.0) 778 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:19 vm00 bash[20709]: audit 2026-03-10T10:20:19.399297+0000 mon.a (mon.0) 778 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:19 vm00 bash[20709]: audit 2026-03-10T10:20:19.433933+0000 mon.a (mon.0) 779 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:19 vm00 bash[20709]: audit 2026-03-10T10:20:19.433933+0000 mon.a (mon.0) 779 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:19 vm00 bash[20709]: audit 2026-03-10T10:20:19.470731+0000 mon.a (mon.0) 780 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:20.166 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:19 vm00 bash[20709]: audit 2026-03-10T10:20:19.470731+0000 mon.a (mon.0) 780 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:22.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:21 vm09 bash[20916]: cluster 2026-03-10T10:20:19.800821+0000 mgr.a (mgr.14150) 257 : cluster [DBG] pgmap v169: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:22.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:21 vm09 bash[20916]: cluster 2026-03-10T10:20:19.800821+0000 mgr.a (mgr.14150) 257 : cluster [DBG] pgmap v169: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:22.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:21 vm03 bash[21274]: cluster 2026-03-10T10:20:19.800821+0000 mgr.a (mgr.14150) 257 : cluster [DBG] pgmap v169: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:22.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:21 vm03 bash[21274]: cluster 2026-03-10T10:20:19.800821+0000 mgr.a (mgr.14150) 257 : cluster [DBG] pgmap v169: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:22.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:21 vm00 bash[20709]: cluster 2026-03-10T10:20:19.800821+0000 mgr.a (mgr.14150) 257 : cluster [DBG] pgmap v169: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:22.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:21 vm00 bash[20709]: cluster 2026-03-10T10:20:19.800821+0000 mgr.a (mgr.14150) 257 : cluster [DBG] pgmap v169: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:23.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:22 vm03 bash[21274]: cluster 2026-03-10T10:20:21.801048+0000 mgr.a (mgr.14150) 258 : cluster [DBG] pgmap v170: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:23.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:22 vm03 bash[21274]: cluster 2026-03-10T10:20:21.801048+0000 mgr.a (mgr.14150) 258 : cluster [DBG] pgmap v170: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:23.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:22 vm00 bash[20709]: cluster 2026-03-10T10:20:21.801048+0000 mgr.a (mgr.14150) 258 : cluster [DBG] pgmap v170: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:23.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:22 vm00 bash[20709]: cluster 2026-03-10T10:20:21.801048+0000 mgr.a (mgr.14150) 258 : cluster [DBG] pgmap v170: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:23.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:22 vm09 bash[20916]: cluster 2026-03-10T10:20:21.801048+0000 mgr.a (mgr.14150) 258 : cluster [DBG] pgmap v170: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:23.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:22 vm09 bash[20916]: cluster 2026-03-10T10:20:21.801048+0000 mgr.a (mgr.14150) 258 : cluster [DBG] pgmap v170: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:25.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:25 vm00 bash[20709]: cluster 2026-03-10T10:20:23.801419+0000 mgr.a (mgr.14150) 259 : cluster [DBG] pgmap v171: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:25.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:25 vm00 bash[20709]: cluster 2026-03-10T10:20:23.801419+0000 mgr.a (mgr.14150) 259 : cluster [DBG] pgmap v171: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:25.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:25 vm00 bash[20709]: audit 2026-03-10T10:20:24.391784+0000 mon.a (mon.0) 781 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:25.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:25 vm00 bash[20709]: audit 2026-03-10T10:20:24.391784+0000 mon.a (mon.0) 781 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:25.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:25 vm00 bash[20709]: audit 2026-03-10T10:20:24.400872+0000 mon.a (mon.0) 782 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:25.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:25 vm00 bash[20709]: audit 2026-03-10T10:20:24.400872+0000 mon.a (mon.0) 782 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:25.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:25 vm00 bash[20709]: audit 2026-03-10T10:20:24.482184+0000 mon.a (mon.0) 783 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:25.666 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:25 vm00 bash[20709]: audit 2026-03-10T10:20:24.482184+0000 mon.a (mon.0) 783 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:25.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:25 vm09 bash[20916]: cluster 2026-03-10T10:20:23.801419+0000 mgr.a (mgr.14150) 259 : cluster [DBG] pgmap v171: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:25.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:25 vm09 bash[20916]: cluster 2026-03-10T10:20:23.801419+0000 mgr.a (mgr.14150) 259 : cluster [DBG] pgmap v171: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:25.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:25 vm09 bash[20916]: audit 2026-03-10T10:20:24.391784+0000 mon.a (mon.0) 781 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:25.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:25 vm09 bash[20916]: audit 2026-03-10T10:20:24.391784+0000 mon.a (mon.0) 781 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:25.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:25 vm09 bash[20916]: audit 2026-03-10T10:20:24.400872+0000 mon.a (mon.0) 782 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:25.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:25 vm09 bash[20916]: audit 2026-03-10T10:20:24.400872+0000 mon.a (mon.0) 782 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:25.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:25 vm09 bash[20916]: audit 2026-03-10T10:20:24.482184+0000 mon.a (mon.0) 783 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:25.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:25 vm09 bash[20916]: audit 2026-03-10T10:20:24.482184+0000 mon.a (mon.0) 783 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:25.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:25 vm03 bash[21274]: cluster 2026-03-10T10:20:23.801419+0000 mgr.a (mgr.14150) 259 : cluster [DBG] pgmap v171: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:25.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:25 vm03 bash[21274]: cluster 2026-03-10T10:20:23.801419+0000 mgr.a (mgr.14150) 259 : cluster [DBG] pgmap v171: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:25.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:25 vm03 bash[21274]: audit 2026-03-10T10:20:24.391784+0000 mon.a (mon.0) 781 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:25.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:25 vm03 bash[21274]: audit 2026-03-10T10:20:24.391784+0000 mon.a (mon.0) 781 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:25.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:25 vm03 bash[21274]: audit 2026-03-10T10:20:24.400872+0000 mon.a (mon.0) 782 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:25.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:25 vm03 bash[21274]: audit 2026-03-10T10:20:24.400872+0000 mon.a (mon.0) 782 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:25.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:25 vm03 bash[21274]: audit 2026-03-10T10:20:24.482184+0000 mon.a (mon.0) 783 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:25.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:25 vm03 bash[21274]: audit 2026-03-10T10:20:24.482184+0000 mon.a (mon.0) 783 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:27.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:27 vm09 bash[20916]: cluster 2026-03-10T10:20:25.801661+0000 mgr.a (mgr.14150) 260 : cluster [DBG] pgmap v172: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:27.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:27 vm09 bash[20916]: cluster 2026-03-10T10:20:25.801661+0000 mgr.a (mgr.14150) 260 : cluster [DBG] pgmap v172: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:27.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:27 vm03 bash[21274]: cluster 2026-03-10T10:20:25.801661+0000 mgr.a (mgr.14150) 260 : cluster [DBG] pgmap v172: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:27.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:27 vm03 bash[21274]: cluster 2026-03-10T10:20:25.801661+0000 mgr.a (mgr.14150) 260 : cluster [DBG] pgmap v172: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:27.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:27 vm00 bash[20709]: cluster 2026-03-10T10:20:25.801661+0000 mgr.a (mgr.14150) 260 : cluster [DBG] pgmap v172: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:27.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:27 vm00 bash[20709]: cluster 2026-03-10T10:20:25.801661+0000 mgr.a (mgr.14150) 260 : cluster [DBG] pgmap v172: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:29.736 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:29 vm09 bash[20916]: cluster 2026-03-10T10:20:27.801916+0000 mgr.a (mgr.14150) 261 : cluster [DBG] pgmap v173: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:29.737 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:29 vm09 bash[20916]: cluster 2026-03-10T10:20:27.801916+0000 mgr.a (mgr.14150) 261 : cluster [DBG] pgmap v173: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:29.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:29 vm03 bash[21274]: cluster 2026-03-10T10:20:27.801916+0000 mgr.a (mgr.14150) 261 : cluster [DBG] pgmap v173: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:29.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:29 vm03 bash[21274]: cluster 2026-03-10T10:20:27.801916+0000 mgr.a (mgr.14150) 261 : cluster [DBG] pgmap v173: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:29.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:29 vm00 bash[20709]: cluster 2026-03-10T10:20:27.801916+0000 mgr.a (mgr.14150) 261 : cluster [DBG] pgmap v173: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:29.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:29 vm00 bash[20709]: cluster 2026-03-10T10:20:27.801916+0000 mgr.a (mgr.14150) 261 : cluster [DBG] pgmap v173: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:31.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:31 vm03 bash[21274]: cluster 2026-03-10T10:20:29.802185+0000 mgr.a (mgr.14150) 262 : cluster [DBG] pgmap v174: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:31.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:31 vm03 bash[21274]: cluster 2026-03-10T10:20:29.802185+0000 mgr.a (mgr.14150) 262 : cluster [DBG] pgmap v174: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:31.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:31 vm00 bash[20709]: cluster 2026-03-10T10:20:29.802185+0000 mgr.a (mgr.14150) 262 : cluster [DBG] pgmap v174: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:31.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:31 vm00 bash[20709]: cluster 2026-03-10T10:20:29.802185+0000 mgr.a (mgr.14150) 262 : cluster [DBG] pgmap v174: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:32.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:31 vm09 bash[20916]: cluster 2026-03-10T10:20:29.802185+0000 mgr.a (mgr.14150) 262 : cluster [DBG] pgmap v174: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:32.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:31 vm09 bash[20916]: cluster 2026-03-10T10:20:29.802185+0000 mgr.a (mgr.14150) 262 : cluster [DBG] pgmap v174: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:33.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:33 vm03 bash[21274]: cluster 2026-03-10T10:20:31.802426+0000 mgr.a (mgr.14150) 263 : cluster [DBG] pgmap v175: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:33.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:33 vm03 bash[21274]: cluster 2026-03-10T10:20:31.802426+0000 mgr.a (mgr.14150) 263 : cluster [DBG] pgmap v175: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:33.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:33 vm00 bash[20709]: cluster 2026-03-10T10:20:31.802426+0000 mgr.a (mgr.14150) 263 : cluster [DBG] pgmap v175: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:33.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:33 vm00 bash[20709]: cluster 2026-03-10T10:20:31.802426+0000 mgr.a (mgr.14150) 263 : cluster [DBG] pgmap v175: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:34.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:33 vm09 bash[20916]: cluster 2026-03-10T10:20:31.802426+0000 mgr.a (mgr.14150) 263 : cluster [DBG] pgmap v175: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:34.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:33 vm09 bash[20916]: cluster 2026-03-10T10:20:31.802426+0000 mgr.a (mgr.14150) 263 : cluster [DBG] pgmap v175: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:35.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:35 vm00 bash[20709]: cluster 2026-03-10T10:20:33.802650+0000 mgr.a (mgr.14150) 264 : cluster [DBG] pgmap v176: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:35.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:35 vm00 bash[20709]: cluster 2026-03-10T10:20:33.802650+0000 mgr.a (mgr.14150) 264 : cluster [DBG] pgmap v176: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:36.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:35 vm09 bash[20916]: cluster 2026-03-10T10:20:33.802650+0000 mgr.a (mgr.14150) 264 : cluster [DBG] pgmap v176: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:36.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:35 vm09 bash[20916]: cluster 2026-03-10T10:20:33.802650+0000 mgr.a (mgr.14150) 264 : cluster [DBG] pgmap v176: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:36.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:35 vm03 bash[21274]: cluster 2026-03-10T10:20:33.802650+0000 mgr.a (mgr.14150) 264 : cluster [DBG] pgmap v176: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:36.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:35 vm03 bash[21274]: cluster 2026-03-10T10:20:33.802650+0000 mgr.a (mgr.14150) 264 : cluster [DBG] pgmap v176: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:37.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:37 vm00 bash[20709]: cluster 2026-03-10T10:20:35.802925+0000 mgr.a (mgr.14150) 265 : cluster [DBG] pgmap v177: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:37.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:37 vm00 bash[20709]: cluster 2026-03-10T10:20:35.802925+0000 mgr.a (mgr.14150) 265 : cluster [DBG] pgmap v177: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:38.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:37 vm09 bash[20916]: cluster 2026-03-10T10:20:35.802925+0000 mgr.a (mgr.14150) 265 : cluster [DBG] pgmap v177: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:38.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:37 vm09 bash[20916]: cluster 2026-03-10T10:20:35.802925+0000 mgr.a (mgr.14150) 265 : cluster [DBG] pgmap v177: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:38.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:37 vm03 bash[21274]: cluster 2026-03-10T10:20:35.802925+0000 mgr.a (mgr.14150) 265 : cluster [DBG] pgmap v177: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:38.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:37 vm03 bash[21274]: cluster 2026-03-10T10:20:35.802925+0000 mgr.a (mgr.14150) 265 : cluster [DBG] pgmap v177: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:39.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:39 vm03 bash[21274]: cluster 2026-03-10T10:20:37.803259+0000 mgr.a (mgr.14150) 266 : cluster [DBG] pgmap v178: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:39.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:39 vm03 bash[21274]: cluster 2026-03-10T10:20:37.803259+0000 mgr.a (mgr.14150) 266 : cluster [DBG] pgmap v178: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:39.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:39 vm03 bash[21274]: audit 2026-03-10T10:20:38.174768+0000 mon.a (mon.0) 784 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:39.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:39 vm03 bash[21274]: audit 2026-03-10T10:20:38.174768+0000 mon.a (mon.0) 784 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:39.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:39 vm03 bash[21274]: audit 2026-03-10T10:20:38.180228+0000 mon.a (mon.0) 785 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:39.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:39 vm03 bash[21274]: audit 2026-03-10T10:20:38.180228+0000 mon.a (mon.0) 785 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:39.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:39 vm03 bash[21274]: audit 2026-03-10T10:20:38.186547+0000 mon.a (mon.0) 786 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:39.331 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:39 vm03 bash[21274]: audit 2026-03-10T10:20:38.186547+0000 mon.a (mon.0) 786 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:39.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:39 vm09 bash[20916]: cluster 2026-03-10T10:20:37.803259+0000 mgr.a (mgr.14150) 266 : cluster [DBG] pgmap v178: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:39.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:39 vm09 bash[20916]: cluster 2026-03-10T10:20:37.803259+0000 mgr.a (mgr.14150) 266 : cluster [DBG] pgmap v178: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:39.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:39 vm09 bash[20916]: audit 2026-03-10T10:20:38.174768+0000 mon.a (mon.0) 784 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:39.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:39 vm09 bash[20916]: audit 2026-03-10T10:20:38.174768+0000 mon.a (mon.0) 784 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:39.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:39 vm09 bash[20916]: audit 2026-03-10T10:20:38.180228+0000 mon.a (mon.0) 785 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:39.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:39 vm09 bash[20916]: audit 2026-03-10T10:20:38.180228+0000 mon.a (mon.0) 785 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:39.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:39 vm09 bash[20916]: audit 2026-03-10T10:20:38.186547+0000 mon.a (mon.0) 786 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:39.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:39 vm09 bash[20916]: audit 2026-03-10T10:20:38.186547+0000 mon.a (mon.0) 786 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:39.552 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:39 vm00 bash[20709]: cluster 2026-03-10T10:20:37.803259+0000 mgr.a (mgr.14150) 266 : cluster [DBG] pgmap v178: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:39.552 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:39 vm00 bash[20709]: cluster 2026-03-10T10:20:37.803259+0000 mgr.a (mgr.14150) 266 : cluster [DBG] pgmap v178: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:39.552 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:39 vm00 bash[20709]: audit 2026-03-10T10:20:38.174768+0000 mon.a (mon.0) 784 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:39.552 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:39 vm00 bash[20709]: audit 2026-03-10T10:20:38.174768+0000 mon.a (mon.0) 784 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:39.552 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:39 vm00 bash[20709]: audit 2026-03-10T10:20:38.180228+0000 mon.a (mon.0) 785 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:39.552 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:39 vm00 bash[20709]: audit 2026-03-10T10:20:38.180228+0000 mon.a (mon.0) 785 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:39.552 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:39 vm00 bash[20709]: audit 2026-03-10T10:20:38.186547+0000 mon.a (mon.0) 786 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:39.552 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:39 vm00 bash[20709]: audit 2026-03-10T10:20:38.186547+0000 mon.a (mon.0) 786 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:40.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:40 vm03 bash[21274]: audit 2026-03-10T10:20:39.544369+0000 mon.a (mon.0) 787 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:40.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:40 vm03 bash[21274]: audit 2026-03-10T10:20:39.544369+0000 mon.a (mon.0) 787 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:40.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:40 vm03 bash[21274]: audit 2026-03-10T10:20:39.548939+0000 mon.a (mon.0) 788 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:40.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:40 vm03 bash[21274]: audit 2026-03-10T10:20:39.548939+0000 mon.a (mon.0) 788 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:40.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:40 vm03 bash[21274]: audit 2026-03-10T10:20:39.552777+0000 mon.a (mon.0) 789 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:40.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:40 vm03 bash[21274]: audit 2026-03-10T10:20:39.552777+0000 mon.a (mon.0) 789 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:40.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:40 vm00 bash[20709]: audit 2026-03-10T10:20:39.544369+0000 mon.a (mon.0) 787 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:40.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:40 vm00 bash[20709]: audit 2026-03-10T10:20:39.544369+0000 mon.a (mon.0) 787 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:40.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:40 vm00 bash[20709]: audit 2026-03-10T10:20:39.548939+0000 mon.a (mon.0) 788 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:40.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:40 vm00 bash[20709]: audit 2026-03-10T10:20:39.548939+0000 mon.a (mon.0) 788 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:40.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:40 vm00 bash[20709]: audit 2026-03-10T10:20:39.552777+0000 mon.a (mon.0) 789 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:40.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:40 vm00 bash[20709]: audit 2026-03-10T10:20:39.552777+0000 mon.a (mon.0) 789 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:41.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:40 vm09 bash[20916]: audit 2026-03-10T10:20:39.544369+0000 mon.a (mon.0) 787 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:41.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:40 vm09 bash[20916]: audit 2026-03-10T10:20:39.544369+0000 mon.a (mon.0) 787 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:41.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:40 vm09 bash[20916]: audit 2026-03-10T10:20:39.548939+0000 mon.a (mon.0) 788 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:41.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:40 vm09 bash[20916]: audit 2026-03-10T10:20:39.548939+0000 mon.a (mon.0) 788 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:41.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:40 vm09 bash[20916]: audit 2026-03-10T10:20:39.552777+0000 mon.a (mon.0) 789 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:41.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:40 vm09 bash[20916]: audit 2026-03-10T10:20:39.552777+0000 mon.a (mon.0) 789 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:41.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:41 vm03 bash[21274]: cluster 2026-03-10T10:20:39.803546+0000 mgr.a (mgr.14150) 267 : cluster [DBG] pgmap v179: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:41.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:41 vm03 bash[21274]: cluster 2026-03-10T10:20:39.803546+0000 mgr.a (mgr.14150) 267 : cluster [DBG] pgmap v179: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:41.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:41 vm00 bash[20709]: cluster 2026-03-10T10:20:39.803546+0000 mgr.a (mgr.14150) 267 : cluster [DBG] pgmap v179: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:41.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:41 vm00 bash[20709]: cluster 2026-03-10T10:20:39.803546+0000 mgr.a (mgr.14150) 267 : cluster [DBG] pgmap v179: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:42.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:41 vm09 bash[20916]: cluster 2026-03-10T10:20:39.803546+0000 mgr.a (mgr.14150) 267 : cluster [DBG] pgmap v179: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:42.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:41 vm09 bash[20916]: cluster 2026-03-10T10:20:39.803546+0000 mgr.a (mgr.14150) 267 : cluster [DBG] pgmap v179: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:43.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:43 vm03 bash[21274]: cluster 2026-03-10T10:20:41.803784+0000 mgr.a (mgr.14150) 268 : cluster [DBG] pgmap v180: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:43.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:43 vm03 bash[21274]: cluster 2026-03-10T10:20:41.803784+0000 mgr.a (mgr.14150) 268 : cluster [DBG] pgmap v180: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:43.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:43 vm00 bash[20709]: cluster 2026-03-10T10:20:41.803784+0000 mgr.a (mgr.14150) 268 : cluster [DBG] pgmap v180: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:43.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:43 vm00 bash[20709]: cluster 2026-03-10T10:20:41.803784+0000 mgr.a (mgr.14150) 268 : cluster [DBG] pgmap v180: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:44.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:43 vm09 bash[20916]: cluster 2026-03-10T10:20:41.803784+0000 mgr.a (mgr.14150) 268 : cluster [DBG] pgmap v180: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:44.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:43 vm09 bash[20916]: cluster 2026-03-10T10:20:41.803784+0000 mgr.a (mgr.14150) 268 : cluster [DBG] pgmap v180: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:45.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:45 vm03 bash[21274]: cluster 2026-03-10T10:20:43.804043+0000 mgr.a (mgr.14150) 269 : cluster [DBG] pgmap v181: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:45 vm03 bash[21274]: cluster 2026-03-10T10:20:43.804043+0000 mgr.a (mgr.14150) 269 : cluster [DBG] pgmap v181: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:45 vm03 bash[21274]: audit 2026-03-10T10:20:44.541735+0000 mon.a (mon.0) 790 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:45 vm03 bash[21274]: audit 2026-03-10T10:20:44.541735+0000 mon.a (mon.0) 790 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:45 vm03 bash[21274]: audit 2026-03-10T10:20:44.545192+0000 mon.a (mon.0) 791 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:45 vm03 bash[21274]: audit 2026-03-10T10:20:44.545192+0000 mon.a (mon.0) 791 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:45 vm03 bash[21274]: audit 2026-03-10T10:20:44.548488+0000 mon.a (mon.0) 792 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:45.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:45 vm03 bash[21274]: audit 2026-03-10T10:20:44.548488+0000 mon.a (mon.0) 792 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:45.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:45 vm00 bash[20709]: cluster 2026-03-10T10:20:43.804043+0000 mgr.a (mgr.14150) 269 : cluster [DBG] pgmap v181: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:45.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:45 vm00 bash[20709]: cluster 2026-03-10T10:20:43.804043+0000 mgr.a (mgr.14150) 269 : cluster [DBG] pgmap v181: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:45.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:45 vm00 bash[20709]: audit 2026-03-10T10:20:44.541735+0000 mon.a (mon.0) 790 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:45.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:45 vm00 bash[20709]: audit 2026-03-10T10:20:44.541735+0000 mon.a (mon.0) 790 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:45.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:45 vm00 bash[20709]: audit 2026-03-10T10:20:44.545192+0000 mon.a (mon.0) 791 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:45.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:45 vm00 bash[20709]: audit 2026-03-10T10:20:44.545192+0000 mon.a (mon.0) 791 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:45.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:45 vm00 bash[20709]: audit 2026-03-10T10:20:44.548488+0000 mon.a (mon.0) 792 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:45.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:45 vm00 bash[20709]: audit 2026-03-10T10:20:44.548488+0000 mon.a (mon.0) 792 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:46.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:45 vm09 bash[20916]: cluster 2026-03-10T10:20:43.804043+0000 mgr.a (mgr.14150) 269 : cluster [DBG] pgmap v181: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:46.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:45 vm09 bash[20916]: cluster 2026-03-10T10:20:43.804043+0000 mgr.a (mgr.14150) 269 : cluster [DBG] pgmap v181: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:46.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:45 vm09 bash[20916]: audit 2026-03-10T10:20:44.541735+0000 mon.a (mon.0) 790 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:46.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:45 vm09 bash[20916]: audit 2026-03-10T10:20:44.541735+0000 mon.a (mon.0) 790 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:46.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:45 vm09 bash[20916]: audit 2026-03-10T10:20:44.545192+0000 mon.a (mon.0) 791 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:46.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:45 vm09 bash[20916]: audit 2026-03-10T10:20:44.545192+0000 mon.a (mon.0) 791 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:46.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:45 vm09 bash[20916]: audit 2026-03-10T10:20:44.548488+0000 mon.a (mon.0) 792 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:46.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:45 vm09 bash[20916]: audit 2026-03-10T10:20:44.548488+0000 mon.a (mon.0) 792 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:47.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:47 vm00 bash[20709]: cluster 2026-03-10T10:20:45.804251+0000 mgr.a (mgr.14150) 270 : cluster [DBG] pgmap v182: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:47.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:47 vm00 bash[20709]: cluster 2026-03-10T10:20:45.804251+0000 mgr.a (mgr.14150) 270 : cluster [DBG] pgmap v182: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:47 vm09 bash[20916]: cluster 2026-03-10T10:20:45.804251+0000 mgr.a (mgr.14150) 270 : cluster [DBG] pgmap v182: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:48.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:47 vm09 bash[20916]: cluster 2026-03-10T10:20:45.804251+0000 mgr.a (mgr.14150) 270 : cluster [DBG] pgmap v182: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:48.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:47 vm03 bash[21274]: cluster 2026-03-10T10:20:45.804251+0000 mgr.a (mgr.14150) 270 : cluster [DBG] pgmap v182: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:48.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:47 vm03 bash[21274]: cluster 2026-03-10T10:20:45.804251+0000 mgr.a (mgr.14150) 270 : cluster [DBG] pgmap v182: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:49.872 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:49 vm09 bash[20916]: cluster 2026-03-10T10:20:47.804542+0000 mgr.a (mgr.14150) 271 : cluster [DBG] pgmap v183: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:49.872 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:49 vm09 bash[20916]: cluster 2026-03-10T10:20:47.804542+0000 mgr.a (mgr.14150) 271 : cluster [DBG] pgmap v183: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:49.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:49 vm00 bash[20709]: cluster 2026-03-10T10:20:47.804542+0000 mgr.a (mgr.14150) 271 : cluster [DBG] pgmap v183: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:49.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:49 vm00 bash[20709]: cluster 2026-03-10T10:20:47.804542+0000 mgr.a (mgr.14150) 271 : cluster [DBG] pgmap v183: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:50.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:49 vm03 bash[21274]: cluster 2026-03-10T10:20:47.804542+0000 mgr.a (mgr.14150) 271 : cluster [DBG] pgmap v183: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:50.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:49 vm03 bash[21274]: cluster 2026-03-10T10:20:47.804542+0000 mgr.a (mgr.14150) 271 : cluster [DBG] pgmap v183: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:51.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:51 vm00 bash[20709]: cluster 2026-03-10T10:20:49.804786+0000 mgr.a (mgr.14150) 272 : cluster [DBG] pgmap v184: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:51.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:51 vm00 bash[20709]: cluster 2026-03-10T10:20:49.804786+0000 mgr.a (mgr.14150) 272 : cluster [DBG] pgmap v184: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:52.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:51 vm09 bash[20916]: cluster 2026-03-10T10:20:49.804786+0000 mgr.a (mgr.14150) 272 : cluster [DBG] pgmap v184: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:52.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:51 vm09 bash[20916]: cluster 2026-03-10T10:20:49.804786+0000 mgr.a (mgr.14150) 272 : cluster [DBG] pgmap v184: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:52.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:51 vm03 bash[21274]: cluster 2026-03-10T10:20:49.804786+0000 mgr.a (mgr.14150) 272 : cluster [DBG] pgmap v184: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:52.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:51 vm03 bash[21274]: cluster 2026-03-10T10:20:49.804786+0000 mgr.a (mgr.14150) 272 : cluster [DBG] pgmap v184: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:53.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:53 vm00 bash[20709]: cluster 2026-03-10T10:20:51.805086+0000 mgr.a (mgr.14150) 273 : cluster [DBG] pgmap v185: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:53.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:53 vm00 bash[20709]: cluster 2026-03-10T10:20:51.805086+0000 mgr.a (mgr.14150) 273 : cluster [DBG] pgmap v185: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:54.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:53 vm09 bash[20916]: cluster 2026-03-10T10:20:51.805086+0000 mgr.a (mgr.14150) 273 : cluster [DBG] pgmap v185: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:54.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:53 vm09 bash[20916]: cluster 2026-03-10T10:20:51.805086+0000 mgr.a (mgr.14150) 273 : cluster [DBG] pgmap v185: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:54.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:53 vm03 bash[21274]: cluster 2026-03-10T10:20:51.805086+0000 mgr.a (mgr.14150) 273 : cluster [DBG] pgmap v185: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:54.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:53 vm03 bash[21274]: cluster 2026-03-10T10:20:51.805086+0000 mgr.a (mgr.14150) 273 : cluster [DBG] pgmap v185: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:55.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:55 vm00 bash[20709]: cluster 2026-03-10T10:20:53.805305+0000 mgr.a (mgr.14150) 274 : cluster [DBG] pgmap v186: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:55.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:55 vm00 bash[20709]: cluster 2026-03-10T10:20:53.805305+0000 mgr.a (mgr.14150) 274 : cluster [DBG] pgmap v186: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:56.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:55 vm09 bash[20916]: cluster 2026-03-10T10:20:53.805305+0000 mgr.a (mgr.14150) 274 : cluster [DBG] pgmap v186: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:56.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:55 vm09 bash[20916]: cluster 2026-03-10T10:20:53.805305+0000 mgr.a (mgr.14150) 274 : cluster [DBG] pgmap v186: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:56.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:55 vm03 bash[21274]: cluster 2026-03-10T10:20:53.805305+0000 mgr.a (mgr.14150) 274 : cluster [DBG] pgmap v186: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:56.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:55 vm03 bash[21274]: cluster 2026-03-10T10:20:53.805305+0000 mgr.a (mgr.14150) 274 : cluster [DBG] pgmap v186: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:57.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:57 vm00 bash[20709]: cluster 2026-03-10T10:20:55.805530+0000 mgr.a (mgr.14150) 275 : cluster [DBG] pgmap v187: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:57.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:57 vm00 bash[20709]: cluster 2026-03-10T10:20:55.805530+0000 mgr.a (mgr.14150) 275 : cluster [DBG] pgmap v187: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:58.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:57 vm09 bash[20916]: cluster 2026-03-10T10:20:55.805530+0000 mgr.a (mgr.14150) 275 : cluster [DBG] pgmap v187: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:58.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:57 vm09 bash[20916]: cluster 2026-03-10T10:20:55.805530+0000 mgr.a (mgr.14150) 275 : cluster [DBG] pgmap v187: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:58.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:57 vm03 bash[21274]: cluster 2026-03-10T10:20:55.805530+0000 mgr.a (mgr.14150) 275 : cluster [DBG] pgmap v187: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:58.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:57 vm03 bash[21274]: cluster 2026-03-10T10:20:55.805530+0000 mgr.a (mgr.14150) 275 : cluster [DBG] pgmap v187: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:59.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:59 vm09 bash[20916]: cluster 2026-03-10T10:20:57.805827+0000 mgr.a (mgr.14150) 276 : cluster [DBG] pgmap v188: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:59.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:59 vm09 bash[20916]: cluster 2026-03-10T10:20:57.805827+0000 mgr.a (mgr.14150) 276 : cluster [DBG] pgmap v188: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:59.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:59 vm09 bash[20916]: audit 2026-03-10T10:20:58.248001+0000 mon.a (mon.0) 793 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:59.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:59 vm09 bash[20916]: audit 2026-03-10T10:20:58.248001+0000 mon.a (mon.0) 793 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:59.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:59 vm09 bash[20916]: audit 2026-03-10T10:20:58.252177+0000 mon.a (mon.0) 794 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:59.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:59 vm09 bash[20916]: audit 2026-03-10T10:20:58.252177+0000 mon.a (mon.0) 794 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:59.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:59 vm09 bash[20916]: audit 2026-03-10T10:20:58.256225+0000 mon.a (mon.0) 795 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:59.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:20:59 vm09 bash[20916]: audit 2026-03-10T10:20:58.256225+0000 mon.a (mon.0) 795 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:59 vm03 bash[21274]: cluster 2026-03-10T10:20:57.805827+0000 mgr.a (mgr.14150) 276 : cluster [DBG] pgmap v188: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:59 vm03 bash[21274]: cluster 2026-03-10T10:20:57.805827+0000 mgr.a (mgr.14150) 276 : cluster [DBG] pgmap v188: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:59 vm03 bash[21274]: audit 2026-03-10T10:20:58.248001+0000 mon.a (mon.0) 793 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:59 vm03 bash[21274]: audit 2026-03-10T10:20:58.248001+0000 mon.a (mon.0) 793 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:59 vm03 bash[21274]: audit 2026-03-10T10:20:58.252177+0000 mon.a (mon.0) 794 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:59 vm03 bash[21274]: audit 2026-03-10T10:20:58.252177+0000 mon.a (mon.0) 794 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:59 vm03 bash[21274]: audit 2026-03-10T10:20:58.256225+0000 mon.a (mon.0) 795 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:59.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:20:59 vm03 bash[21274]: audit 2026-03-10T10:20:58.256225+0000 mon.a (mon.0) 795 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:59.621 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:59 vm00 bash[20709]: cluster 2026-03-10T10:20:57.805827+0000 mgr.a (mgr.14150) 276 : cluster [DBG] pgmap v188: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:59.621 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:59 vm00 bash[20709]: cluster 2026-03-10T10:20:57.805827+0000 mgr.a (mgr.14150) 276 : cluster [DBG] pgmap v188: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:20:59.621 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:59 vm00 bash[20709]: audit 2026-03-10T10:20:58.248001+0000 mon.a (mon.0) 793 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:59.621 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:59 vm00 bash[20709]: audit 2026-03-10T10:20:58.248001+0000 mon.a (mon.0) 793 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:59.621 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:59 vm00 bash[20709]: audit 2026-03-10T10:20:58.252177+0000 mon.a (mon.0) 794 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:59.621 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:59 vm00 bash[20709]: audit 2026-03-10T10:20:58.252177+0000 mon.a (mon.0) 794 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:59.621 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:59 vm00 bash[20709]: audit 2026-03-10T10:20:58.256225+0000 mon.a (mon.0) 795 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:20:59.621 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:20:59 vm00 bash[20709]: audit 2026-03-10T10:20:58.256225+0000 mon.a (mon.0) 795 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:00.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:00 vm03 bash[21274]: audit 2026-03-10T10:20:59.612997+0000 mon.a (mon.0) 796 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:00.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:00 vm03 bash[21274]: audit 2026-03-10T10:20:59.612997+0000 mon.a (mon.0) 796 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:00.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:00 vm03 bash[21274]: audit 2026-03-10T10:20:59.617124+0000 mon.a (mon.0) 797 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:00.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:00 vm03 bash[21274]: audit 2026-03-10T10:20:59.617124+0000 mon.a (mon.0) 797 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:00.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:00 vm03 bash[21274]: audit 2026-03-10T10:20:59.621674+0000 mon.a (mon.0) 798 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:00.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:00 vm03 bash[21274]: audit 2026-03-10T10:20:59.621674+0000 mon.a (mon.0) 798 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:00.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:00 vm00 bash[20709]: audit 2026-03-10T10:20:59.612997+0000 mon.a (mon.0) 796 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:00.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:00 vm00 bash[20709]: audit 2026-03-10T10:20:59.612997+0000 mon.a (mon.0) 796 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:00.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:00 vm00 bash[20709]: audit 2026-03-10T10:20:59.617124+0000 mon.a (mon.0) 797 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:00.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:00 vm00 bash[20709]: audit 2026-03-10T10:20:59.617124+0000 mon.a (mon.0) 797 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:00.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:00 vm00 bash[20709]: audit 2026-03-10T10:20:59.621674+0000 mon.a (mon.0) 798 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:00.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:00 vm00 bash[20709]: audit 2026-03-10T10:20:59.621674+0000 mon.a (mon.0) 798 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:01.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:00 vm09 bash[20916]: audit 2026-03-10T10:20:59.612997+0000 mon.a (mon.0) 796 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:01.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:00 vm09 bash[20916]: audit 2026-03-10T10:20:59.612997+0000 mon.a (mon.0) 796 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:01.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:00 vm09 bash[20916]: audit 2026-03-10T10:20:59.617124+0000 mon.a (mon.0) 797 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:01.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:00 vm09 bash[20916]: audit 2026-03-10T10:20:59.617124+0000 mon.a (mon.0) 797 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:01.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:00 vm09 bash[20916]: audit 2026-03-10T10:20:59.621674+0000 mon.a (mon.0) 798 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:01.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:00 vm09 bash[20916]: audit 2026-03-10T10:20:59.621674+0000 mon.a (mon.0) 798 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:01.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:01 vm00 bash[20709]: cluster 2026-03-10T10:20:59.806074+0000 mgr.a (mgr.14150) 277 : cluster [DBG] pgmap v189: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:01.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:01 vm00 bash[20709]: cluster 2026-03-10T10:20:59.806074+0000 mgr.a (mgr.14150) 277 : cluster [DBG] pgmap v189: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:02.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:01 vm09 bash[20916]: cluster 2026-03-10T10:20:59.806074+0000 mgr.a (mgr.14150) 277 : cluster [DBG] pgmap v189: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:02.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:01 vm09 bash[20916]: cluster 2026-03-10T10:20:59.806074+0000 mgr.a (mgr.14150) 277 : cluster [DBG] pgmap v189: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:02.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:01 vm03 bash[21274]: cluster 2026-03-10T10:20:59.806074+0000 mgr.a (mgr.14150) 277 : cluster [DBG] pgmap v189: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:02.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:01 vm03 bash[21274]: cluster 2026-03-10T10:20:59.806074+0000 mgr.a (mgr.14150) 277 : cluster [DBG] pgmap v189: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:04.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:03 vm00 bash[20709]: cluster 2026-03-10T10:21:01.806341+0000 mgr.a (mgr.14150) 278 : cluster [DBG] pgmap v190: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:04.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:03 vm00 bash[20709]: cluster 2026-03-10T10:21:01.806341+0000 mgr.a (mgr.14150) 278 : cluster [DBG] pgmap v190: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:04.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:03 vm03 bash[21274]: cluster 2026-03-10T10:21:01.806341+0000 mgr.a (mgr.14150) 278 : cluster [DBG] pgmap v190: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:04.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:03 vm03 bash[21274]: cluster 2026-03-10T10:21:01.806341+0000 mgr.a (mgr.14150) 278 : cluster [DBG] pgmap v190: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:04.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:04 vm09 bash[20916]: cluster 2026-03-10T10:21:01.806341+0000 mgr.a (mgr.14150) 278 : cluster [DBG] pgmap v190: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:04.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:04 vm09 bash[20916]: cluster 2026-03-10T10:21:01.806341+0000 mgr.a (mgr.14150) 278 : cluster [DBG] pgmap v190: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:05.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:05 vm00 bash[20709]: cluster 2026-03-10T10:21:03.806637+0000 mgr.a (mgr.14150) 279 : cluster [DBG] pgmap v191: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:05.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:05 vm00 bash[20709]: cluster 2026-03-10T10:21:03.806637+0000 mgr.a (mgr.14150) 279 : cluster [DBG] pgmap v191: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:05.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:05 vm00 bash[20709]: audit 2026-03-10T10:21:04.616153+0000 mon.a (mon.0) 799 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:05.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:05 vm00 bash[20709]: audit 2026-03-10T10:21:04.616153+0000 mon.a (mon.0) 799 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:05.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:05 vm00 bash[20709]: audit 2026-03-10T10:21:04.619845+0000 mon.a (mon.0) 800 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:05.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:05 vm00 bash[20709]: audit 2026-03-10T10:21:04.619845+0000 mon.a (mon.0) 800 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:05.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:05 vm00 bash[20709]: audit 2026-03-10T10:21:04.624011+0000 mon.a (mon.0) 801 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:05.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:05 vm00 bash[20709]: audit 2026-03-10T10:21:04.624011+0000 mon.a (mon.0) 801 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:06.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:05 vm09 bash[20916]: cluster 2026-03-10T10:21:03.806637+0000 mgr.a (mgr.14150) 279 : cluster [DBG] pgmap v191: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:06.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:05 vm09 bash[20916]: cluster 2026-03-10T10:21:03.806637+0000 mgr.a (mgr.14150) 279 : cluster [DBG] pgmap v191: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:06.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:05 vm09 bash[20916]: audit 2026-03-10T10:21:04.616153+0000 mon.a (mon.0) 799 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:06.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:05 vm09 bash[20916]: audit 2026-03-10T10:21:04.616153+0000 mon.a (mon.0) 799 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:06.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:05 vm09 bash[20916]: audit 2026-03-10T10:21:04.619845+0000 mon.a (mon.0) 800 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:06.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:05 vm09 bash[20916]: audit 2026-03-10T10:21:04.619845+0000 mon.a (mon.0) 800 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:06.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:05 vm09 bash[20916]: audit 2026-03-10T10:21:04.624011+0000 mon.a (mon.0) 801 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:06.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:05 vm09 bash[20916]: audit 2026-03-10T10:21:04.624011+0000 mon.a (mon.0) 801 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:06.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:05 vm03 bash[21274]: cluster 2026-03-10T10:21:03.806637+0000 mgr.a (mgr.14150) 279 : cluster [DBG] pgmap v191: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:06.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:05 vm03 bash[21274]: cluster 2026-03-10T10:21:03.806637+0000 mgr.a (mgr.14150) 279 : cluster [DBG] pgmap v191: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:06.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:05 vm03 bash[21274]: audit 2026-03-10T10:21:04.616153+0000 mon.a (mon.0) 799 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:06.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:05 vm03 bash[21274]: audit 2026-03-10T10:21:04.616153+0000 mon.a (mon.0) 799 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:06.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:05 vm03 bash[21274]: audit 2026-03-10T10:21:04.619845+0000 mon.a (mon.0) 800 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:06.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:05 vm03 bash[21274]: audit 2026-03-10T10:21:04.619845+0000 mon.a (mon.0) 800 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:06.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:05 vm03 bash[21274]: audit 2026-03-10T10:21:04.624011+0000 mon.a (mon.0) 801 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:06.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:05 vm03 bash[21274]: audit 2026-03-10T10:21:04.624011+0000 mon.a (mon.0) 801 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:08.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:07 vm00 bash[20709]: cluster 2026-03-10T10:21:05.806846+0000 mgr.a (mgr.14150) 280 : cluster [DBG] pgmap v192: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:08.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:07 vm00 bash[20709]: cluster 2026-03-10T10:21:05.806846+0000 mgr.a (mgr.14150) 280 : cluster [DBG] pgmap v192: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:08.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:08 vm09 bash[20916]: cluster 2026-03-10T10:21:05.806846+0000 mgr.a (mgr.14150) 280 : cluster [DBG] pgmap v192: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:08.267 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:08 vm09 bash[20916]: cluster 2026-03-10T10:21:05.806846+0000 mgr.a (mgr.14150) 280 : cluster [DBG] pgmap v192: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:08.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:08 vm03 bash[21274]: cluster 2026-03-10T10:21:05.806846+0000 mgr.a (mgr.14150) 280 : cluster [DBG] pgmap v192: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:08.340 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:08 vm03 bash[21274]: cluster 2026-03-10T10:21:05.806846+0000 mgr.a (mgr.14150) 280 : cluster [DBG] pgmap v192: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:09.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:09 vm03 bash[21274]: cluster 2026-03-10T10:21:07.807106+0000 mgr.a (mgr.14150) 281 : cluster [DBG] pgmap v193: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:09.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:09 vm03 bash[21274]: cluster 2026-03-10T10:21:07.807106+0000 mgr.a (mgr.14150) 281 : cluster [DBG] pgmap v193: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:09.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:09 vm00 bash[20709]: cluster 2026-03-10T10:21:07.807106+0000 mgr.a (mgr.14150) 281 : cluster [DBG] pgmap v193: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:09.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:09 vm00 bash[20709]: cluster 2026-03-10T10:21:07.807106+0000 mgr.a (mgr.14150) 281 : cluster [DBG] pgmap v193: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:09.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:09 vm09 bash[20916]: cluster 2026-03-10T10:21:07.807106+0000 mgr.a (mgr.14150) 281 : cluster [DBG] pgmap v193: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:09.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:09 vm09 bash[20916]: cluster 2026-03-10T10:21:07.807106+0000 mgr.a (mgr.14150) 281 : cluster [DBG] pgmap v193: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:11.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:11 vm00 bash[20709]: cluster 2026-03-10T10:21:09.807386+0000 mgr.a (mgr.14150) 282 : cluster [DBG] pgmap v194: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:11.665 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:11 vm00 bash[20709]: cluster 2026-03-10T10:21:09.807386+0000 mgr.a (mgr.14150) 282 : cluster [DBG] pgmap v194: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:11.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:11 vm09 bash[20916]: cluster 2026-03-10T10:21:09.807386+0000 mgr.a (mgr.14150) 282 : cluster [DBG] pgmap v194: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:11.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:11 vm09 bash[20916]: cluster 2026-03-10T10:21:09.807386+0000 mgr.a (mgr.14150) 282 : cluster [DBG] pgmap v194: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:11.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:11 vm03 bash[21274]: cluster 2026-03-10T10:21:09.807386+0000 mgr.a (mgr.14150) 282 : cluster [DBG] pgmap v194: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:11.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:11 vm03 bash[21274]: cluster 2026-03-10T10:21:09.807386+0000 mgr.a (mgr.14150) 282 : cluster [DBG] pgmap v194: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:13.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:13 vm09 bash[20916]: cluster 2026-03-10T10:21:11.807627+0000 mgr.a (mgr.14150) 283 : cluster [DBG] pgmap v195: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:13.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:13 vm09 bash[20916]: cluster 2026-03-10T10:21:11.807627+0000 mgr.a (mgr.14150) 283 : cluster [DBG] pgmap v195: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:13.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:13 vm03 bash[21274]: cluster 2026-03-10T10:21:11.807627+0000 mgr.a (mgr.14150) 283 : cluster [DBG] pgmap v195: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:13.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:13 vm03 bash[21274]: cluster 2026-03-10T10:21:11.807627+0000 mgr.a (mgr.14150) 283 : cluster [DBG] pgmap v195: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:13.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:13 vm00 bash[20709]: cluster 2026-03-10T10:21:11.807627+0000 mgr.a (mgr.14150) 283 : cluster [DBG] pgmap v195: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:13.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:13 vm00 bash[20709]: cluster 2026-03-10T10:21:11.807627+0000 mgr.a (mgr.14150) 283 : cluster [DBG] pgmap v195: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:15.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:15 vm09 bash[20916]: cluster 2026-03-10T10:21:13.807872+0000 mgr.a (mgr.14150) 284 : cluster [DBG] pgmap v196: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:15.758 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:15 vm09 bash[20916]: cluster 2026-03-10T10:21:13.807872+0000 mgr.a (mgr.14150) 284 : cluster [DBG] pgmap v196: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:15.813 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:15 vm03 bash[21274]: cluster 2026-03-10T10:21:13.807872+0000 mgr.a (mgr.14150) 284 : cluster [DBG] pgmap v196: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:15.813 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:15 vm03 bash[21274]: cluster 2026-03-10T10:21:13.807872+0000 mgr.a (mgr.14150) 284 : cluster [DBG] pgmap v196: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:15.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:15 vm00 bash[20709]: cluster 2026-03-10T10:21:13.807872+0000 mgr.a (mgr.14150) 284 : cluster [DBG] pgmap v196: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:15.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:15 vm00 bash[20709]: cluster 2026-03-10T10:21:13.807872+0000 mgr.a (mgr.14150) 284 : cluster [DBG] pgmap v196: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:17.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:17 vm03 bash[21274]: cluster 2026-03-10T10:21:15.808268+0000 mgr.a (mgr.14150) 285 : cluster [DBG] pgmap v197: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:17.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:17 vm03 bash[21274]: cluster 2026-03-10T10:21:15.808268+0000 mgr.a (mgr.14150) 285 : cluster [DBG] pgmap v197: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:17.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:17 vm00 bash[20709]: cluster 2026-03-10T10:21:15.808268+0000 mgr.a (mgr.14150) 285 : cluster [DBG] pgmap v197: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:17.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:17 vm00 bash[20709]: cluster 2026-03-10T10:21:15.808268+0000 mgr.a (mgr.14150) 285 : cluster [DBG] pgmap v197: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:18.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:17 vm09 bash[20916]: cluster 2026-03-10T10:21:15.808268+0000 mgr.a (mgr.14150) 285 : cluster [DBG] pgmap v197: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:18.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:17 vm09 bash[20916]: cluster 2026-03-10T10:21:15.808268+0000 mgr.a (mgr.14150) 285 : cluster [DBG] pgmap v197: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:18.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:18 vm03 bash[21274]: audit 2026-03-10T10:21:18.064222+0000 mon.a (mon.0) 802 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:21:18.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:18 vm03 bash[21274]: audit 2026-03-10T10:21:18.064222+0000 mon.a (mon.0) 802 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:21:18.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:18 vm03 bash[21274]: audit 2026-03-10T10:21:18.065522+0000 mon.a (mon.0) 803 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:21:18.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:18 vm03 bash[21274]: audit 2026-03-10T10:21:18.065522+0000 mon.a (mon.0) 803 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:21:18.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:18 vm03 bash[21274]: audit 2026-03-10T10:21:18.065922+0000 mon.a (mon.0) 804 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:21:18.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:18 vm03 bash[21274]: audit 2026-03-10T10:21:18.065922+0000 mon.a (mon.0) 804 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:21:18.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:18 vm03 bash[21274]: audit 2026-03-10T10:21:18.070337+0000 mon.a (mon.0) 805 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:18.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:18 vm03 bash[21274]: audit 2026-03-10T10:21:18.070337+0000 mon.a (mon.0) 805 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:18.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:18 vm03 bash[21274]: audit 2026-03-10T10:21:18.314578+0000 mon.a (mon.0) 806 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:18.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:18 vm03 bash[21274]: audit 2026-03-10T10:21:18.314578+0000 mon.a (mon.0) 806 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:18.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:18 vm03 bash[21274]: audit 2026-03-10T10:21:18.318540+0000 mon.a (mon.0) 807 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:18.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:18 vm03 bash[21274]: audit 2026-03-10T10:21:18.318540+0000 mon.a (mon.0) 807 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:18.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:18 vm03 bash[21274]: audit 2026-03-10T10:21:18.322505+0000 mon.a (mon.0) 808 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:18.581 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:18 vm03 bash[21274]: audit 2026-03-10T10:21:18.322505+0000 mon.a (mon.0) 808 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:18.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:18 vm00 bash[20709]: audit 2026-03-10T10:21:18.064222+0000 mon.a (mon.0) 802 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:21:18.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:18 vm00 bash[20709]: audit 2026-03-10T10:21:18.064222+0000 mon.a (mon.0) 802 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:21:18.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:18 vm00 bash[20709]: audit 2026-03-10T10:21:18.065522+0000 mon.a (mon.0) 803 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:21:18.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:18 vm00 bash[20709]: audit 2026-03-10T10:21:18.065522+0000 mon.a (mon.0) 803 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:21:18.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:18 vm00 bash[20709]: audit 2026-03-10T10:21:18.065922+0000 mon.a (mon.0) 804 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:21:18.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:18 vm00 bash[20709]: audit 2026-03-10T10:21:18.065922+0000 mon.a (mon.0) 804 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:21:18.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:18 vm00 bash[20709]: audit 2026-03-10T10:21:18.070337+0000 mon.a (mon.0) 805 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:18.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:18 vm00 bash[20709]: audit 2026-03-10T10:21:18.070337+0000 mon.a (mon.0) 805 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:18.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:18 vm00 bash[20709]: audit 2026-03-10T10:21:18.314578+0000 mon.a (mon.0) 806 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:18.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:18 vm00 bash[20709]: audit 2026-03-10T10:21:18.314578+0000 mon.a (mon.0) 806 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:18.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:18 vm00 bash[20709]: audit 2026-03-10T10:21:18.318540+0000 mon.a (mon.0) 807 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:18.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:18 vm00 bash[20709]: audit 2026-03-10T10:21:18.318540+0000 mon.a (mon.0) 807 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:18.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:18 vm00 bash[20709]: audit 2026-03-10T10:21:18.322505+0000 mon.a (mon.0) 808 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:18.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:18 vm00 bash[20709]: audit 2026-03-10T10:21:18.322505+0000 mon.a (mon.0) 808 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:19.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:18 vm09 bash[20916]: audit 2026-03-10T10:21:18.064222+0000 mon.a (mon.0) 802 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:21:19.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:18 vm09 bash[20916]: audit 2026-03-10T10:21:18.064222+0000 mon.a (mon.0) 802 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:21:19.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:18 vm09 bash[20916]: audit 2026-03-10T10:21:18.065522+0000 mon.a (mon.0) 803 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:21:19.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:18 vm09 bash[20916]: audit 2026-03-10T10:21:18.065522+0000 mon.a (mon.0) 803 : audit [DBG] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:21:19.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:18 vm09 bash[20916]: audit 2026-03-10T10:21:18.065922+0000 mon.a (mon.0) 804 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:21:19.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:18 vm09 bash[20916]: audit 2026-03-10T10:21:18.065922+0000 mon.a (mon.0) 804 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:21:19.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:18 vm09 bash[20916]: audit 2026-03-10T10:21:18.070337+0000 mon.a (mon.0) 805 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:19.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:18 vm09 bash[20916]: audit 2026-03-10T10:21:18.070337+0000 mon.a (mon.0) 805 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:19.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:18 vm09 bash[20916]: audit 2026-03-10T10:21:18.314578+0000 mon.a (mon.0) 806 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:19.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:18 vm09 bash[20916]: audit 2026-03-10T10:21:18.314578+0000 mon.a (mon.0) 806 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:19.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:18 vm09 bash[20916]: audit 2026-03-10T10:21:18.318540+0000 mon.a (mon.0) 807 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:19.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:18 vm09 bash[20916]: audit 2026-03-10T10:21:18.318540+0000 mon.a (mon.0) 807 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:19.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:18 vm09 bash[20916]: audit 2026-03-10T10:21:18.322505+0000 mon.a (mon.0) 808 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:19.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:18 vm09 bash[20916]: audit 2026-03-10T10:21:18.322505+0000 mon.a (mon.0) 808 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:19.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:19 vm03 bash[21274]: cluster 2026-03-10T10:21:17.808596+0000 mgr.a (mgr.14150) 286 : cluster [DBG] pgmap v198: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:19.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:19 vm03 bash[21274]: cluster 2026-03-10T10:21:17.808596+0000 mgr.a (mgr.14150) 286 : cluster [DBG] pgmap v198: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:19.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:19 vm00 bash[20709]: cluster 2026-03-10T10:21:17.808596+0000 mgr.a (mgr.14150) 286 : cluster [DBG] pgmap v198: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:19.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:19 vm00 bash[20709]: cluster 2026-03-10T10:21:17.808596+0000 mgr.a (mgr.14150) 286 : cluster [DBG] pgmap v198: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:20.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:19 vm09 bash[20916]: cluster 2026-03-10T10:21:17.808596+0000 mgr.a (mgr.14150) 286 : cluster [DBG] pgmap v198: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:20.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:19 vm09 bash[20916]: cluster 2026-03-10T10:21:17.808596+0000 mgr.a (mgr.14150) 286 : cluster [DBG] pgmap v198: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:21.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:20 vm03 bash[21274]: cluster 2026-03-10T10:21:19.808820+0000 mgr.a (mgr.14150) 287 : cluster [DBG] pgmap v199: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:21.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:20 vm03 bash[21274]: cluster 2026-03-10T10:21:19.808820+0000 mgr.a (mgr.14150) 287 : cluster [DBG] pgmap v199: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:21.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:20 vm03 bash[21274]: audit 2026-03-10T10:21:19.837169+0000 mon.a (mon.0) 809 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:21.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:20 vm03 bash[21274]: audit 2026-03-10T10:21:19.837169+0000 mon.a (mon.0) 809 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:21.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:20 vm03 bash[21274]: audit 2026-03-10T10:21:19.849599+0000 mon.a (mon.0) 810 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:21.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:20 vm03 bash[21274]: audit 2026-03-10T10:21:19.849599+0000 mon.a (mon.0) 810 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:21.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:20 vm03 bash[21274]: audit 2026-03-10T10:21:19.952332+0000 mon.a (mon.0) 811 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:21.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:20 vm03 bash[21274]: audit 2026-03-10T10:21:19.952332+0000 mon.a (mon.0) 811 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:21.090 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:20 vm00 bash[20709]: cluster 2026-03-10T10:21:19.808820+0000 mgr.a (mgr.14150) 287 : cluster [DBG] pgmap v199: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:21.090 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:20 vm00 bash[20709]: cluster 2026-03-10T10:21:19.808820+0000 mgr.a (mgr.14150) 287 : cluster [DBG] pgmap v199: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:21.090 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:20 vm00 bash[20709]: audit 2026-03-10T10:21:19.837169+0000 mon.a (mon.0) 809 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:21.090 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:20 vm00 bash[20709]: audit 2026-03-10T10:21:19.837169+0000 mon.a (mon.0) 809 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:21.090 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:20 vm00 bash[20709]: audit 2026-03-10T10:21:19.849599+0000 mon.a (mon.0) 810 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:21.090 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:20 vm00 bash[20709]: audit 2026-03-10T10:21:19.849599+0000 mon.a (mon.0) 810 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:21.090 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:20 vm00 bash[20709]: audit 2026-03-10T10:21:19.952332+0000 mon.a (mon.0) 811 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:21.090 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:20 vm00 bash[20709]: audit 2026-03-10T10:21:19.952332+0000 mon.a (mon.0) 811 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:21.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:20 vm09 bash[20916]: cluster 2026-03-10T10:21:19.808820+0000 mgr.a (mgr.14150) 287 : cluster [DBG] pgmap v199: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:21.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:20 vm09 bash[20916]: cluster 2026-03-10T10:21:19.808820+0000 mgr.a (mgr.14150) 287 : cluster [DBG] pgmap v199: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:21.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:20 vm09 bash[20916]: audit 2026-03-10T10:21:19.837169+0000 mon.a (mon.0) 809 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:21.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:20 vm09 bash[20916]: audit 2026-03-10T10:21:19.837169+0000 mon.a (mon.0) 809 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:21.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:20 vm09 bash[20916]: audit 2026-03-10T10:21:19.849599+0000 mon.a (mon.0) 810 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:21.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:20 vm09 bash[20916]: audit 2026-03-10T10:21:19.849599+0000 mon.a (mon.0) 810 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:21.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:20 vm09 bash[20916]: audit 2026-03-10T10:21:19.952332+0000 mon.a (mon.0) 811 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:21.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:20 vm09 bash[20916]: audit 2026-03-10T10:21:19.952332+0000 mon.a (mon.0) 811 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:23.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:22 vm00 bash[20709]: cluster 2026-03-10T10:21:21.809138+0000 mgr.a (mgr.14150) 288 : cluster [DBG] pgmap v200: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:23.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:22 vm00 bash[20709]: cluster 2026-03-10T10:21:21.809138+0000 mgr.a (mgr.14150) 288 : cluster [DBG] pgmap v200: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:23.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:22 vm09 bash[20916]: cluster 2026-03-10T10:21:21.809138+0000 mgr.a (mgr.14150) 288 : cluster [DBG] pgmap v200: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:23.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:22 vm09 bash[20916]: cluster 2026-03-10T10:21:21.809138+0000 mgr.a (mgr.14150) 288 : cluster [DBG] pgmap v200: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:23.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:22 vm03 bash[21274]: cluster 2026-03-10T10:21:21.809138+0000 mgr.a (mgr.14150) 288 : cluster [DBG] pgmap v200: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:23.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:22 vm03 bash[21274]: cluster 2026-03-10T10:21:21.809138+0000 mgr.a (mgr.14150) 288 : cluster [DBG] pgmap v200: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:26.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:25 vm09 bash[20916]: cluster 2026-03-10T10:21:23.809452+0000 mgr.a (mgr.14150) 289 : cluster [DBG] pgmap v201: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:26.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:25 vm09 bash[20916]: cluster 2026-03-10T10:21:23.809452+0000 mgr.a (mgr.14150) 289 : cluster [DBG] pgmap v201: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:26.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:25 vm09 bash[20916]: audit 2026-03-10T10:21:24.682607+0000 mon.a (mon.0) 812 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:26.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:25 vm09 bash[20916]: audit 2026-03-10T10:21:24.682607+0000 mon.a (mon.0) 812 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:26.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:25 vm09 bash[20916]: audit 2026-03-10T10:21:24.687605+0000 mon.a (mon.0) 813 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:26.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:25 vm09 bash[20916]: audit 2026-03-10T10:21:24.687605+0000 mon.a (mon.0) 813 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:26.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:25 vm09 bash[20916]: audit 2026-03-10T10:21:24.692179+0000 mon.a (mon.0) 814 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:26.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:25 vm09 bash[20916]: audit 2026-03-10T10:21:24.692179+0000 mon.a (mon.0) 814 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:26.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:25 vm03 bash[21274]: cluster 2026-03-10T10:21:23.809452+0000 mgr.a (mgr.14150) 289 : cluster [DBG] pgmap v201: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:26.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:25 vm03 bash[21274]: cluster 2026-03-10T10:21:23.809452+0000 mgr.a (mgr.14150) 289 : cluster [DBG] pgmap v201: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:26.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:25 vm03 bash[21274]: audit 2026-03-10T10:21:24.682607+0000 mon.a (mon.0) 812 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:26.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:25 vm03 bash[21274]: audit 2026-03-10T10:21:24.682607+0000 mon.a (mon.0) 812 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:26.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:25 vm03 bash[21274]: audit 2026-03-10T10:21:24.687605+0000 mon.a (mon.0) 813 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:26.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:25 vm03 bash[21274]: audit 2026-03-10T10:21:24.687605+0000 mon.a (mon.0) 813 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:26.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:25 vm03 bash[21274]: audit 2026-03-10T10:21:24.692179+0000 mon.a (mon.0) 814 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:26.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:25 vm03 bash[21274]: audit 2026-03-10T10:21:24.692179+0000 mon.a (mon.0) 814 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:26.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:25 vm00 bash[20709]: cluster 2026-03-10T10:21:23.809452+0000 mgr.a (mgr.14150) 289 : cluster [DBG] pgmap v201: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:26.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:25 vm00 bash[20709]: cluster 2026-03-10T10:21:23.809452+0000 mgr.a (mgr.14150) 289 : cluster [DBG] pgmap v201: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:26.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:25 vm00 bash[20709]: audit 2026-03-10T10:21:24.682607+0000 mon.a (mon.0) 812 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:26.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:25 vm00 bash[20709]: audit 2026-03-10T10:21:24.682607+0000 mon.a (mon.0) 812 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:26.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:25 vm00 bash[20709]: audit 2026-03-10T10:21:24.687605+0000 mon.a (mon.0) 813 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:26.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:25 vm00 bash[20709]: audit 2026-03-10T10:21:24.687605+0000 mon.a (mon.0) 813 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:26.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:25 vm00 bash[20709]: audit 2026-03-10T10:21:24.692179+0000 mon.a (mon.0) 814 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:26.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:25 vm00 bash[20709]: audit 2026-03-10T10:21:24.692179+0000 mon.a (mon.0) 814 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:28.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:27 vm09 bash[20916]: cluster 2026-03-10T10:21:25.809674+0000 mgr.a (mgr.14150) 290 : cluster [DBG] pgmap v202: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:28.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:27 vm09 bash[20916]: cluster 2026-03-10T10:21:25.809674+0000 mgr.a (mgr.14150) 290 : cluster [DBG] pgmap v202: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:28.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:27 vm03 bash[21274]: cluster 2026-03-10T10:21:25.809674+0000 mgr.a (mgr.14150) 290 : cluster [DBG] pgmap v202: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:28.082 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:27 vm03 bash[21274]: cluster 2026-03-10T10:21:25.809674+0000 mgr.a (mgr.14150) 290 : cluster [DBG] pgmap v202: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:28.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:27 vm00 bash[20709]: cluster 2026-03-10T10:21:25.809674+0000 mgr.a (mgr.14150) 290 : cluster [DBG] pgmap v202: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:28.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:27 vm00 bash[20709]: cluster 2026-03-10T10:21:25.809674+0000 mgr.a (mgr.14150) 290 : cluster [DBG] pgmap v202: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:30.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:29 vm09 bash[20916]: cluster 2026-03-10T10:21:27.809938+0000 mgr.a (mgr.14150) 291 : cluster [DBG] pgmap v203: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:30.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:29 vm09 bash[20916]: cluster 2026-03-10T10:21:27.809938+0000 mgr.a (mgr.14150) 291 : cluster [DBG] pgmap v203: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:30.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:29 vm03 bash[21274]: cluster 2026-03-10T10:21:27.809938+0000 mgr.a (mgr.14150) 291 : cluster [DBG] pgmap v203: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:30.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:29 vm03 bash[21274]: cluster 2026-03-10T10:21:27.809938+0000 mgr.a (mgr.14150) 291 : cluster [DBG] pgmap v203: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:30.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:29 vm00 bash[20709]: cluster 2026-03-10T10:21:27.809938+0000 mgr.a (mgr.14150) 291 : cluster [DBG] pgmap v203: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:30.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:29 vm00 bash[20709]: cluster 2026-03-10T10:21:27.809938+0000 mgr.a (mgr.14150) 291 : cluster [DBG] pgmap v203: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:32.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:31 vm09 bash[20916]: cluster 2026-03-10T10:21:29.810159+0000 mgr.a (mgr.14150) 292 : cluster [DBG] pgmap v204: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:32.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:31 vm09 bash[20916]: cluster 2026-03-10T10:21:29.810159+0000 mgr.a (mgr.14150) 292 : cluster [DBG] pgmap v204: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:32.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:31 vm03 bash[21274]: cluster 2026-03-10T10:21:29.810159+0000 mgr.a (mgr.14150) 292 : cluster [DBG] pgmap v204: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:32.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:31 vm03 bash[21274]: cluster 2026-03-10T10:21:29.810159+0000 mgr.a (mgr.14150) 292 : cluster [DBG] pgmap v204: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:32.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:31 vm00 bash[20709]: cluster 2026-03-10T10:21:29.810159+0000 mgr.a (mgr.14150) 292 : cluster [DBG] pgmap v204: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:32.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:31 vm00 bash[20709]: cluster 2026-03-10T10:21:29.810159+0000 mgr.a (mgr.14150) 292 : cluster [DBG] pgmap v204: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:34.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:33 vm03 bash[21274]: cluster 2026-03-10T10:21:31.810375+0000 mgr.a (mgr.14150) 293 : cluster [DBG] pgmap v205: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:34.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:33 vm03 bash[21274]: cluster 2026-03-10T10:21:31.810375+0000 mgr.a (mgr.14150) 293 : cluster [DBG] pgmap v205: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:34.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:33 vm00 bash[20709]: cluster 2026-03-10T10:21:31.810375+0000 mgr.a (mgr.14150) 293 : cluster [DBG] pgmap v205: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:34.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:33 vm00 bash[20709]: cluster 2026-03-10T10:21:31.810375+0000 mgr.a (mgr.14150) 293 : cluster [DBG] pgmap v205: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:34.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:33 vm09 bash[20916]: cluster 2026-03-10T10:21:31.810375+0000 mgr.a (mgr.14150) 293 : cluster [DBG] pgmap v205: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:34.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:33 vm09 bash[20916]: cluster 2026-03-10T10:21:31.810375+0000 mgr.a (mgr.14150) 293 : cluster [DBG] pgmap v205: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:35.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:34 vm00 bash[20709]: cluster 2026-03-10T10:21:33.810589+0000 mgr.a (mgr.14150) 294 : cluster [DBG] pgmap v206: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:35.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:34 vm00 bash[20709]: cluster 2026-03-10T10:21:33.810589+0000 mgr.a (mgr.14150) 294 : cluster [DBG] pgmap v206: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:35.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:34 vm09 bash[20916]: cluster 2026-03-10T10:21:33.810589+0000 mgr.a (mgr.14150) 294 : cluster [DBG] pgmap v206: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:35.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:34 vm09 bash[20916]: cluster 2026-03-10T10:21:33.810589+0000 mgr.a (mgr.14150) 294 : cluster [DBG] pgmap v206: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:35.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:34 vm03 bash[21274]: cluster 2026-03-10T10:21:33.810589+0000 mgr.a (mgr.14150) 294 : cluster [DBG] pgmap v206: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:35.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:34 vm03 bash[21274]: cluster 2026-03-10T10:21:33.810589+0000 mgr.a (mgr.14150) 294 : cluster [DBG] pgmap v206: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:37.107 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:36 vm00 bash[20709]: cluster 2026-03-10T10:21:35.810870+0000 mgr.a (mgr.14150) 295 : cluster [DBG] pgmap v207: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:37.108 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:36 vm00 bash[20709]: cluster 2026-03-10T10:21:35.810870+0000 mgr.a (mgr.14150) 295 : cluster [DBG] pgmap v207: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:37.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:36 vm09 bash[20916]: cluster 2026-03-10T10:21:35.810870+0000 mgr.a (mgr.14150) 295 : cluster [DBG] pgmap v207: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:37.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:36 vm09 bash[20916]: cluster 2026-03-10T10:21:35.810870+0000 mgr.a (mgr.14150) 295 : cluster [DBG] pgmap v207: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:37.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:36 vm03 bash[21274]: cluster 2026-03-10T10:21:35.810870+0000 mgr.a (mgr.14150) 295 : cluster [DBG] pgmap v207: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:37.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:36 vm03 bash[21274]: cluster 2026-03-10T10:21:35.810870+0000 mgr.a (mgr.14150) 295 : cluster [DBG] pgmap v207: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:39.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:39 vm03 bash[21274]: cluster 2026-03-10T10:21:37.811245+0000 mgr.a (mgr.14150) 296 : cluster [DBG] pgmap v208: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:39.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:39 vm03 bash[21274]: cluster 2026-03-10T10:21:37.811245+0000 mgr.a (mgr.14150) 296 : cluster [DBG] pgmap v208: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:39.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:39 vm03 bash[21274]: audit 2026-03-10T10:21:38.527928+0000 mon.a (mon.0) 815 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:39.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:39 vm03 bash[21274]: audit 2026-03-10T10:21:38.527928+0000 mon.a (mon.0) 815 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:39.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:39 vm03 bash[21274]: audit 2026-03-10T10:21:38.657664+0000 mon.a (mon.0) 816 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:39.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:39 vm03 bash[21274]: audit 2026-03-10T10:21:38.657664+0000 mon.a (mon.0) 816 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:39.830 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:39 vm03 bash[21274]: audit 2026-03-10T10:21:38.807764+0000 mon.a (mon.0) 817 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:39.831 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:39 vm03 bash[21274]: audit 2026-03-10T10:21:38.807764+0000 mon.a (mon.0) 817 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:39.915 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:39 vm00 bash[20709]: cluster 2026-03-10T10:21:37.811245+0000 mgr.a (mgr.14150) 296 : cluster [DBG] pgmap v208: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:39.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:39 vm00 bash[20709]: cluster 2026-03-10T10:21:37.811245+0000 mgr.a (mgr.14150) 296 : cluster [DBG] pgmap v208: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:39.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:39 vm00 bash[20709]: audit 2026-03-10T10:21:38.527928+0000 mon.a (mon.0) 815 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:39.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:39 vm00 bash[20709]: audit 2026-03-10T10:21:38.527928+0000 mon.a (mon.0) 815 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:39.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:39 vm00 bash[20709]: audit 2026-03-10T10:21:38.657664+0000 mon.a (mon.0) 816 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:39.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:39 vm00 bash[20709]: audit 2026-03-10T10:21:38.657664+0000 mon.a (mon.0) 816 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:39.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:39 vm00 bash[20709]: audit 2026-03-10T10:21:38.807764+0000 mon.a (mon.0) 817 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:39.916 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:39 vm00 bash[20709]: audit 2026-03-10T10:21:38.807764+0000 mon.a (mon.0) 817 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:40.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:39 vm09 bash[20916]: cluster 2026-03-10T10:21:37.811245+0000 mgr.a (mgr.14150) 296 : cluster [DBG] pgmap v208: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:40.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:39 vm09 bash[20916]: cluster 2026-03-10T10:21:37.811245+0000 mgr.a (mgr.14150) 296 : cluster [DBG] pgmap v208: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:40.007 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:39 vm09 bash[20916]: audit 2026-03-10T10:21:38.527928+0000 mon.a (mon.0) 815 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:40.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:39 vm09 bash[20916]: audit 2026-03-10T10:21:38.527928+0000 mon.a (mon.0) 815 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:40.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:39 vm09 bash[20916]: audit 2026-03-10T10:21:38.657664+0000 mon.a (mon.0) 816 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:40.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:39 vm09 bash[20916]: audit 2026-03-10T10:21:38.657664+0000 mon.a (mon.0) 816 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:40.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:39 vm09 bash[20916]: audit 2026-03-10T10:21:38.807764+0000 mon.a (mon.0) 817 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:40.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:39 vm09 bash[20916]: audit 2026-03-10T10:21:38.807764+0000 mon.a (mon.0) 817 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.415 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:41 vm00 bash[20709]: cluster 2026-03-10T10:21:39.811501+0000 mgr.a (mgr.14150) 297 : cluster [DBG] pgmap v209: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:41.415 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:41 vm00 bash[20709]: cluster 2026-03-10T10:21:39.811501+0000 mgr.a (mgr.14150) 297 : cluster [DBG] pgmap v209: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:41.415 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:41 vm00 bash[20709]: audit 2026-03-10T10:21:40.119221+0000 mon.a (mon.0) 818 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.415 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:41 vm00 bash[20709]: audit 2026-03-10T10:21:40.119221+0000 mon.a (mon.0) 818 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.415 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:41 vm00 bash[20709]: audit 2026-03-10T10:21:40.131677+0000 mon.a (mon.0) 819 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.415 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:41 vm00 bash[20709]: audit 2026-03-10T10:21:40.131677+0000 mon.a (mon.0) 819 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.415 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:41 vm00 bash[20709]: audit 2026-03-10T10:21:40.136331+0000 mon.a (mon.0) 820 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.415 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:41 vm00 bash[20709]: audit 2026-03-10T10:21:40.136331+0000 mon.a (mon.0) 820 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:41 vm09 bash[20916]: cluster 2026-03-10T10:21:39.811501+0000 mgr.a (mgr.14150) 297 : cluster [DBG] pgmap v209: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:41.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:41 vm09 bash[20916]: cluster 2026-03-10T10:21:39.811501+0000 mgr.a (mgr.14150) 297 : cluster [DBG] pgmap v209: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:41.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:41 vm09 bash[20916]: audit 2026-03-10T10:21:40.119221+0000 mon.a (mon.0) 818 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:41 vm09 bash[20916]: audit 2026-03-10T10:21:40.119221+0000 mon.a (mon.0) 818 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:41 vm09 bash[20916]: audit 2026-03-10T10:21:40.131677+0000 mon.a (mon.0) 819 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:41 vm09 bash[20916]: audit 2026-03-10T10:21:40.131677+0000 mon.a (mon.0) 819 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:41 vm09 bash[20916]: audit 2026-03-10T10:21:40.136331+0000 mon.a (mon.0) 820 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:41 vm09 bash[20916]: audit 2026-03-10T10:21:40.136331+0000 mon.a (mon.0) 820 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:41 vm03 bash[21274]: cluster 2026-03-10T10:21:39.811501+0000 mgr.a (mgr.14150) 297 : cluster [DBG] pgmap v209: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:41.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:41 vm03 bash[21274]: cluster 2026-03-10T10:21:39.811501+0000 mgr.a (mgr.14150) 297 : cluster [DBG] pgmap v209: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:41.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:41 vm03 bash[21274]: audit 2026-03-10T10:21:40.119221+0000 mon.a (mon.0) 818 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:41 vm03 bash[21274]: audit 2026-03-10T10:21:40.119221+0000 mon.a (mon.0) 818 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:41 vm03 bash[21274]: audit 2026-03-10T10:21:40.131677+0000 mon.a (mon.0) 819 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:41 vm03 bash[21274]: audit 2026-03-10T10:21:40.131677+0000 mon.a (mon.0) 819 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:41 vm03 bash[21274]: audit 2026-03-10T10:21:40.136331+0000 mon.a (mon.0) 820 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:41.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:41 vm03 bash[21274]: audit 2026-03-10T10:21:40.136331+0000 mon.a (mon.0) 820 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:43.415 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:43 vm00 bash[20709]: cluster 2026-03-10T10:21:41.811720+0000 mgr.a (mgr.14150) 298 : cluster [DBG] pgmap v210: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:43.415 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:43 vm00 bash[20709]: cluster 2026-03-10T10:21:41.811720+0000 mgr.a (mgr.14150) 298 : cluster [DBG] pgmap v210: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:43.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:43 vm09 bash[20916]: cluster 2026-03-10T10:21:41.811720+0000 mgr.a (mgr.14150) 298 : cluster [DBG] pgmap v210: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:43.508 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:43 vm09 bash[20916]: cluster 2026-03-10T10:21:41.811720+0000 mgr.a (mgr.14150) 298 : cluster [DBG] pgmap v210: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:43.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:43 vm03 bash[21274]: cluster 2026-03-10T10:21:41.811720+0000 mgr.a (mgr.14150) 298 : cluster [DBG] pgmap v210: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:43.580 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:43 vm03 bash[21274]: cluster 2026-03-10T10:21:41.811720+0000 mgr.a (mgr.14150) 298 : cluster [DBG] pgmap v210: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:46.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:45 vm09 bash[20916]: cluster 2026-03-10T10:21:43.811962+0000 mgr.a (mgr.14150) 299 : cluster [DBG] pgmap v211: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:46.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:45 vm09 bash[20916]: cluster 2026-03-10T10:21:43.811962+0000 mgr.a (mgr.14150) 299 : cluster [DBG] pgmap v211: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:46.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:45 vm09 bash[20916]: audit 2026-03-10T10:21:44.749545+0000 mon.a (mon.0) 821 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:46.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:45 vm09 bash[20916]: audit 2026-03-10T10:21:44.749545+0000 mon.a (mon.0) 821 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:46.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:45 vm09 bash[20916]: audit 2026-03-10T10:21:44.752952+0000 mon.a (mon.0) 822 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:46.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:45 vm09 bash[20916]: audit 2026-03-10T10:21:44.752952+0000 mon.a (mon.0) 822 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:46.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:45 vm09 bash[20916]: audit 2026-03-10T10:21:44.756429+0000 mon.a (mon.0) 823 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:46.008 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:45 vm09 bash[20916]: audit 2026-03-10T10:21:44.756429+0000 mon.a (mon.0) 823 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:46.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:45 vm03 bash[21274]: cluster 2026-03-10T10:21:43.811962+0000 mgr.a (mgr.14150) 299 : cluster [DBG] pgmap v211: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:46.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:45 vm03 bash[21274]: cluster 2026-03-10T10:21:43.811962+0000 mgr.a (mgr.14150) 299 : cluster [DBG] pgmap v211: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:46.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:45 vm03 bash[21274]: audit 2026-03-10T10:21:44.749545+0000 mon.a (mon.0) 821 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:46.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:45 vm03 bash[21274]: audit 2026-03-10T10:21:44.749545+0000 mon.a (mon.0) 821 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:46.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:45 vm03 bash[21274]: audit 2026-03-10T10:21:44.752952+0000 mon.a (mon.0) 822 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:46.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:45 vm03 bash[21274]: audit 2026-03-10T10:21:44.752952+0000 mon.a (mon.0) 822 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:46.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:45 vm03 bash[21274]: audit 2026-03-10T10:21:44.756429+0000 mon.a (mon.0) 823 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:46.081 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:45 vm03 bash[21274]: audit 2026-03-10T10:21:44.756429+0000 mon.a (mon.0) 823 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:46.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:45 vm00 bash[20709]: cluster 2026-03-10T10:21:43.811962+0000 mgr.a (mgr.14150) 299 : cluster [DBG] pgmap v211: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:46.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:45 vm00 bash[20709]: cluster 2026-03-10T10:21:43.811962+0000 mgr.a (mgr.14150) 299 : cluster [DBG] pgmap v211: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:46.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:45 vm00 bash[20709]: audit 2026-03-10T10:21:44.749545+0000 mon.a (mon.0) 821 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:46.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:45 vm00 bash[20709]: audit 2026-03-10T10:21:44.749545+0000 mon.a (mon.0) 821 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:46.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:45 vm00 bash[20709]: audit 2026-03-10T10:21:44.752952+0000 mon.a (mon.0) 822 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:46.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:45 vm00 bash[20709]: audit 2026-03-10T10:21:44.752952+0000 mon.a (mon.0) 822 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:46.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:45 vm00 bash[20709]: audit 2026-03-10T10:21:44.756429+0000 mon.a (mon.0) 823 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:46.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:45 vm00 bash[20709]: audit 2026-03-10T10:21:44.756429+0000 mon.a (mon.0) 823 : audit [INF] from='mgr.14150 192.168.123.100:0/3658118396' entity='mgr.a' 2026-03-10T10:21:47.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:46 vm03 bash[21274]: cluster 2026-03-10T10:21:45.812205+0000 mgr.a (mgr.14150) 300 : cluster [DBG] pgmap v212: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:47.080 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:46 vm03 bash[21274]: cluster 2026-03-10T10:21:45.812205+0000 mgr.a (mgr.14150) 300 : cluster [DBG] pgmap v212: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:47.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:46 vm00 bash[20709]: cluster 2026-03-10T10:21:45.812205+0000 mgr.a (mgr.14150) 300 : cluster [DBG] pgmap v212: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:47.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:46 vm00 bash[20709]: cluster 2026-03-10T10:21:45.812205+0000 mgr.a (mgr.14150) 300 : cluster [DBG] pgmap v212: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:47.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:46 vm09 bash[20916]: cluster 2026-03-10T10:21:45.812205+0000 mgr.a (mgr.14150) 300 : cluster [DBG] pgmap v212: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:47.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:46 vm09 bash[20916]: cluster 2026-03-10T10:21:45.812205+0000 mgr.a (mgr.14150) 300 : cluster [DBG] pgmap v212: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:48.049 INFO:teuthology.orchestra.run.vm00.stderr:+ ceph mon dump 2026-03-10T10:21:48.231 INFO:teuthology.orchestra.run.vm00.stdout:epoch 6 2026-03-10T10:21:48.231 INFO:teuthology.orchestra.run.vm00.stdout:fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:21:48.231 INFO:teuthology.orchestra.run.vm00.stdout:last_changed 2026-03-10T10:18:58.467065+0000 2026-03-10T10:21:48.231 INFO:teuthology.orchestra.run.vm00.stdout:created 2026-03-10T10:14:44.248116+0000 2026-03-10T10:21:48.231 INFO:teuthology.orchestra.run.vm00.stdout:min_mon_release 19 (squid) 2026-03-10T10:21:48.231 INFO:teuthology.orchestra.run.vm00.stdout:election_strategy: 1 2026-03-10T10:21:48.231 INFO:teuthology.orchestra.run.vm00.stdout:0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a; crush_location {datacenter=a} 2026-03-10T10:21:48.231 INFO:teuthology.orchestra.run.vm00.stdout:1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c; crush_location {datacenter=a,rack=3} 2026-03-10T10:21:48.232 INFO:teuthology.orchestra.run.vm00.stdout:2: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.b; crush_location {datacenter=b,rack=2} 2026-03-10T10:21:48.232 INFO:teuthology.orchestra.run.vm00.stderr:dumped monmap epoch 6 2026-03-10T10:21:48.240 INFO:teuthology.orchestra.run.vm00.stderr:+ ceph mon dump --format json 2026-03-10T10:21:48.422 INFO:teuthology.orchestra.run.vm00.stderr:dumped monmap epoch 6 2026-03-10T10:21:48.422 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-10T10:21:48.422 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":6,"fsid":"9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad","modified":"2026-03-10T10:18:58.467065Z","created":"2026-03-10T10:14:44.248116Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:3300","nonce":0},{"type":"v1","addr":"192.168.123.100:6789","nonce":0}]},"addr":"192.168.123.100:6789/0","public_addr":"192.168.123.100:6789/0","priority":0,"weight":0,"crush_location":"{datacenter=a}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:3300","nonce":0},{"type":"v1","addr":"192.168.123.109:6789","nonce":0}]},"addr":"192.168.123.109:6789/0","public_addr":"192.168.123.109:6789/0","priority":0,"weight":0,"crush_location":"{datacenter=a,rack=3}"},{"rank":2,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:3300","nonce":0},{"type":"v1","addr":"192.168.123.103:6789","nonce":0}]},"addr":"192.168.123.103:6789/0","public_addr":"192.168.123.103:6789/0","priority":0,"weight":0,"crush_location":"{datacenter=b,rack=2}"}],"quorum":[0,1,2]} 2026-03-10T10:21:48.432 INFO:teuthology.orchestra.run.vm00.stderr:+ read monid 2026-03-10T10:21:48.432 INFO:teuthology.orchestra.run.vm00.stderr:+ read crushloc 2026-03-10T10:21:48.432 INFO:teuthology.orchestra.run.vm00.stderr:+ ceph mon dump --format json 2026-03-10T10:21:48.432 INFO:teuthology.orchestra.run.vm00.stderr:+ jq --arg monid a --arg crushloc '{datacenter=a}' -e '.mons | .[] | select(.name == $monid) | .crush_location == $crushloc' 2026-03-10T10:21:48.614 INFO:teuthology.orchestra.run.vm00.stderr:dumped monmap epoch 6 2026-03-10T10:21:48.623 INFO:teuthology.orchestra.run.vm00.stdout:true 2026-03-10T10:21:48.624 INFO:teuthology.orchestra.run.vm00.stderr:+ read monid 2026-03-10T10:21:48.624 INFO:teuthology.orchestra.run.vm00.stderr:+ read crushloc 2026-03-10T10:21:48.624 INFO:teuthology.orchestra.run.vm00.stderr:+ ceph mon dump --format json 2026-03-10T10:21:48.625 INFO:teuthology.orchestra.run.vm00.stderr:+ jq --arg monid b --arg crushloc '{datacenter=b,rack=2}' -e '.mons | .[] | select(.name == $monid) | .crush_location == $crushloc' 2026-03-10T10:21:48.802 INFO:teuthology.orchestra.run.vm00.stderr:dumped monmap epoch 6 2026-03-10T10:21:48.811 INFO:teuthology.orchestra.run.vm00.stdout:true 2026-03-10T10:21:48.811 INFO:teuthology.orchestra.run.vm00.stderr:+ read monid 2026-03-10T10:21:48.812 INFO:teuthology.orchestra.run.vm00.stderr:+ read crushloc 2026-03-10T10:21:48.812 INFO:teuthology.orchestra.run.vm00.stderr:+ ceph mon dump --format json 2026-03-10T10:21:48.812 INFO:teuthology.orchestra.run.vm00.stderr:+ jq --arg monid c --arg crushloc '{datacenter=a,rack=3}' -e '.mons | .[] | select(.name == $monid) | .crush_location == $crushloc' 2026-03-10T10:21:48.993 INFO:teuthology.orchestra.run.vm00.stderr:dumped monmap epoch 6 2026-03-10T10:21:49.003 INFO:teuthology.orchestra.run.vm00.stdout:true 2026-03-10T10:21:49.003 INFO:teuthology.orchestra.run.vm00.stderr:+ read monid 2026-03-10T10:21:49.054 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-10T10:21:49.056 INFO:tasks.cephadm:Teardown begin 2026-03-10T10:21:49.056 DEBUG:teuthology.orchestra.run.vm00:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:21:49.064 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:21:49.071 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:21:49.079 INFO:tasks.cephadm:Disabling cephadm mgr module 2026-03-10T10:21:49.079 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad -- ceph mgr module disable cephadm 2026-03-10T10:21:49.117 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:48 vm00 bash[20709]: cluster 2026-03-10T10:21:47.812436+0000 mgr.a (mgr.14150) 301 : cluster [DBG] pgmap v213: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:49.117 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:48 vm00 bash[20709]: cluster 2026-03-10T10:21:47.812436+0000 mgr.a (mgr.14150) 301 : cluster [DBG] pgmap v213: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:49.118 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:48 vm00 bash[20709]: audit 2026-03-10T10:21:48.233330+0000 mon.a (mon.0) 824 : audit [DBG] from='client.? 192.168.123.100:0/2075275169' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-10T10:21:49.118 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:48 vm00 bash[20709]: audit 2026-03-10T10:21:48.233330+0000 mon.a (mon.0) 824 : audit [DBG] from='client.? 192.168.123.100:0/2075275169' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-10T10:21:49.118 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:48 vm00 bash[20709]: audit 2026-03-10T10:21:48.424028+0000 mon.a (mon.0) 825 : audit [DBG] from='client.? 192.168.123.100:0/3434095270' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:49.118 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:48 vm00 bash[20709]: audit 2026-03-10T10:21:48.424028+0000 mon.a (mon.0) 825 : audit [DBG] from='client.? 192.168.123.100:0/3434095270' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:49.118 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:48 vm00 bash[20709]: audit 2026-03-10T10:21:48.616305+0000 mon.a (mon.0) 826 : audit [DBG] from='client.? 192.168.123.100:0/56446360' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:49.118 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:48 vm00 bash[20709]: audit 2026-03-10T10:21:48.616305+0000 mon.a (mon.0) 826 : audit [DBG] from='client.? 192.168.123.100:0/56446360' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:49.118 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:48 vm00 bash[20709]: audit 2026-03-10T10:21:48.804084+0000 mon.a (mon.0) 827 : audit [DBG] from='client.? 192.168.123.100:0/1464378675' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:49.118 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:48 vm00 bash[20709]: audit 2026-03-10T10:21:48.804084+0000 mon.a (mon.0) 827 : audit [DBG] from='client.? 192.168.123.100:0/1464378675' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:49.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:48 vm09 bash[20916]: cluster 2026-03-10T10:21:47.812436+0000 mgr.a (mgr.14150) 301 : cluster [DBG] pgmap v213: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:49.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:48 vm09 bash[20916]: cluster 2026-03-10T10:21:47.812436+0000 mgr.a (mgr.14150) 301 : cluster [DBG] pgmap v213: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:49.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:48 vm09 bash[20916]: audit 2026-03-10T10:21:48.233330+0000 mon.a (mon.0) 824 : audit [DBG] from='client.? 192.168.123.100:0/2075275169' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-10T10:21:49.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:48 vm09 bash[20916]: audit 2026-03-10T10:21:48.233330+0000 mon.a (mon.0) 824 : audit [DBG] from='client.? 192.168.123.100:0/2075275169' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-10T10:21:49.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:48 vm09 bash[20916]: audit 2026-03-10T10:21:48.424028+0000 mon.a (mon.0) 825 : audit [DBG] from='client.? 192.168.123.100:0/3434095270' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:49.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:48 vm09 bash[20916]: audit 2026-03-10T10:21:48.424028+0000 mon.a (mon.0) 825 : audit [DBG] from='client.? 192.168.123.100:0/3434095270' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:49.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:48 vm09 bash[20916]: audit 2026-03-10T10:21:48.616305+0000 mon.a (mon.0) 826 : audit [DBG] from='client.? 192.168.123.100:0/56446360' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:49.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:48 vm09 bash[20916]: audit 2026-03-10T10:21:48.616305+0000 mon.a (mon.0) 826 : audit [DBG] from='client.? 192.168.123.100:0/56446360' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:49.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:48 vm09 bash[20916]: audit 2026-03-10T10:21:48.804084+0000 mon.a (mon.0) 827 : audit [DBG] from='client.? 192.168.123.100:0/1464378675' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:49.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:48 vm09 bash[20916]: audit 2026-03-10T10:21:48.804084+0000 mon.a (mon.0) 827 : audit [DBG] from='client.? 192.168.123.100:0/1464378675' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:49.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:48 vm03 bash[21274]: cluster 2026-03-10T10:21:47.812436+0000 mgr.a (mgr.14150) 301 : cluster [DBG] pgmap v213: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:49.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:48 vm03 bash[21274]: cluster 2026-03-10T10:21:47.812436+0000 mgr.a (mgr.14150) 301 : cluster [DBG] pgmap v213: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:49.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:48 vm03 bash[21274]: audit 2026-03-10T10:21:48.233330+0000 mon.a (mon.0) 824 : audit [DBG] from='client.? 192.168.123.100:0/2075275169' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-10T10:21:49.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:48 vm03 bash[21274]: audit 2026-03-10T10:21:48.233330+0000 mon.a (mon.0) 824 : audit [DBG] from='client.? 192.168.123.100:0/2075275169' entity='client.admin' cmd=[{"prefix": "mon dump"}]: dispatch 2026-03-10T10:21:49.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:48 vm03 bash[21274]: audit 2026-03-10T10:21:48.424028+0000 mon.a (mon.0) 825 : audit [DBG] from='client.? 192.168.123.100:0/3434095270' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:49.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:48 vm03 bash[21274]: audit 2026-03-10T10:21:48.424028+0000 mon.a (mon.0) 825 : audit [DBG] from='client.? 192.168.123.100:0/3434095270' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:49.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:48 vm03 bash[21274]: audit 2026-03-10T10:21:48.616305+0000 mon.a (mon.0) 826 : audit [DBG] from='client.? 192.168.123.100:0/56446360' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:49.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:48 vm03 bash[21274]: audit 2026-03-10T10:21:48.616305+0000 mon.a (mon.0) 826 : audit [DBG] from='client.? 192.168.123.100:0/56446360' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:49.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:48 vm03 bash[21274]: audit 2026-03-10T10:21:48.804084+0000 mon.a (mon.0) 827 : audit [DBG] from='client.? 192.168.123.100:0/1464378675' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:49.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:48 vm03 bash[21274]: audit 2026-03-10T10:21:48.804084+0000 mon.a (mon.0) 827 : audit [DBG] from='client.? 192.168.123.100:0/1464378675' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:50.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:49 vm00 bash[20709]: audit 2026-03-10T10:21:48.994947+0000 mon.a (mon.0) 828 : audit [DBG] from='client.? 192.168.123.100:0/2202156093' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:50.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:49 vm00 bash[20709]: audit 2026-03-10T10:21:48.994947+0000 mon.a (mon.0) 828 : audit [DBG] from='client.? 192.168.123.100:0/2202156093' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:50.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:49 vm09 bash[20916]: audit 2026-03-10T10:21:48.994947+0000 mon.a (mon.0) 828 : audit [DBG] from='client.? 192.168.123.100:0/2202156093' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:50.258 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:49 vm09 bash[20916]: audit 2026-03-10T10:21:48.994947+0000 mon.a (mon.0) 828 : audit [DBG] from='client.? 192.168.123.100:0/2202156093' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:50.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:49 vm03 bash[21274]: audit 2026-03-10T10:21:48.994947+0000 mon.a (mon.0) 828 : audit [DBG] from='client.? 192.168.123.100:0/2202156093' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:50.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:49 vm03 bash[21274]: audit 2026-03-10T10:21:48.994947+0000 mon.a (mon.0) 828 : audit [DBG] from='client.? 192.168.123.100:0/2202156093' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:21:51.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:50 vm00 bash[20709]: cluster 2026-03-10T10:21:49.812655+0000 mgr.a (mgr.14150) 302 : cluster [DBG] pgmap v214: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:51.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:50 vm00 bash[20709]: cluster 2026-03-10T10:21:49.812655+0000 mgr.a (mgr.14150) 302 : cluster [DBG] pgmap v214: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:51.227 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:50 vm09 bash[20916]: cluster 2026-03-10T10:21:49.812655+0000 mgr.a (mgr.14150) 302 : cluster [DBG] pgmap v214: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:51.227 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:50 vm09 bash[20916]: cluster 2026-03-10T10:21:49.812655+0000 mgr.a (mgr.14150) 302 : cluster [DBG] pgmap v214: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:51.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:50 vm03 bash[21274]: cluster 2026-03-10T10:21:49.812655+0000 mgr.a (mgr.14150) 302 : cluster [DBG] pgmap v214: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:51.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:50 vm03 bash[21274]: cluster 2026-03-10T10:21:49.812655+0000 mgr.a (mgr.14150) 302 : cluster [DBG] pgmap v214: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:53.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:52 vm00 bash[20709]: cluster 2026-03-10T10:21:51.812894+0000 mgr.a (mgr.14150) 303 : cluster [DBG] pgmap v215: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:53.165 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 10 10:21:52 vm00 bash[20709]: cluster 2026-03-10T10:21:51.812894+0000 mgr.a (mgr.14150) 303 : cluster [DBG] pgmap v215: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:53.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:52 vm09 bash[20916]: cluster 2026-03-10T10:21:51.812894+0000 mgr.a (mgr.14150) 303 : cluster [DBG] pgmap v215: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:53.257 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 10:21:52 vm09 bash[20916]: cluster 2026-03-10T10:21:51.812894+0000 mgr.a (mgr.14150) 303 : cluster [DBG] pgmap v215: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:53.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:52 vm03 bash[21274]: cluster 2026-03-10T10:21:51.812894+0000 mgr.a (mgr.14150) 303 : cluster [DBG] pgmap v215: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:53.330 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:52 vm03 bash[21274]: cluster 2026-03-10T10:21:51.812894+0000 mgr.a (mgr.14150) 303 : cluster [DBG] pgmap v215: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T10:21:53.727 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/mon.a/config 2026-03-10T10:21:53.857 INFO:teuthology.orchestra.run.vm00.stderr:2026-03-10T10:21:53.856+0000 7f8b5539e640 -1 auth: error reading file: /etc/ceph/ceph.keyring: bufferlist::read_file(/etc/ceph/ceph.keyring): read error:(21) Is a directory 2026-03-10T10:21:53.857 INFO:teuthology.orchestra.run.vm00.stderr:2026-03-10T10:21:53.856+0000 7f8b5539e640 -1 auth: failed to load /etc/ceph/ceph.keyring: (21) Is a directory 2026-03-10T10:21:53.857 INFO:teuthology.orchestra.run.vm00.stderr:2026-03-10T10:21:53.856+0000 7f8b5539e640 -1 auth: error reading file: /etc/ceph/ceph.keyring: bufferlist::read_file(/etc/ceph/ceph.keyring): read error:(21) Is a directory 2026-03-10T10:21:53.857 INFO:teuthology.orchestra.run.vm00.stderr:2026-03-10T10:21:53.856+0000 7f8b5539e640 -1 auth: failed to load /etc/ceph/ceph.keyring: (21) Is a directory 2026-03-10T10:21:53.857 INFO:teuthology.orchestra.run.vm00.stderr:2026-03-10T10:21:53.856+0000 7f8b5539e640 -1 auth: error reading file: /etc/ceph/ceph.keyring: bufferlist::read_file(/etc/ceph/ceph.keyring): read error:(21) Is a directory 2026-03-10T10:21:53.857 INFO:teuthology.orchestra.run.vm00.stderr:2026-03-10T10:21:53.856+0000 7f8b5539e640 -1 auth: failed to load /etc/ceph/ceph.keyring: (21) Is a directory 2026-03-10T10:21:53.858 INFO:teuthology.orchestra.run.vm00.stderr:2026-03-10T10:21:53.856+0000 7f8b5539e640 -1 monclient: keyring not found 2026-03-10T10:21:53.858 INFO:teuthology.orchestra.run.vm00.stderr:[errno 21] error connecting to the cluster 2026-03-10T10:21:53.903 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:21:53.903 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-10T10:21:53.903 DEBUG:teuthology.orchestra.run.vm00:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-10T10:21:53.947 DEBUG:teuthology.orchestra.run.vm03:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-10T10:21:53.950 DEBUG:teuthology.orchestra.run.vm09:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-10T10:21:53.953 INFO:tasks.cephadm:Stopping all daemons... 2026-03-10T10:21:53.953 INFO:tasks.cephadm.mon.a:Stopping mon.a... 2026-03-10T10:21:53.953 DEBUG:teuthology.orchestra.run.vm00:> sudo systemctl stop ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mon.a 2026-03-10T10:21:54.141 DEBUG:teuthology.orchestra.run.vm00:> sudo pkill -f 'journalctl -f -n 0 -u ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mon.a.service' 2026-03-10T10:21:54.154 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T10:21:54.154 INFO:tasks.cephadm.mon.a:Stopped mon.a 2026-03-10T10:21:54.154 INFO:tasks.cephadm.mon.c:Stopping mon.b... 2026-03-10T10:21:54.154 DEBUG:teuthology.orchestra.run.vm03:> sudo systemctl stop ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mon.b 2026-03-10T10:21:54.417 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:54 vm03 systemd[1]: Stopping Ceph mon.b for 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad... 2026-03-10T10:21:54.417 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:54 vm03 bash[21274]: debug 2026-03-10T10:21:54.206+0000 7fdc3000b640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-10T10:21:54.417 INFO:journalctl@ceph.mon.b.vm03.stdout:Mar 10 10:21:54 vm03 bash[21274]: debug 2026-03-10T10:21:54.206+0000 7fdc3000b640 -1 mon.b@2(peon) e6 *** Got Signal Terminated *** 2026-03-10T10:21:54.460 DEBUG:teuthology.orchestra.run.vm03:> sudo pkill -f 'journalctl -f -n 0 -u ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mon.b.service' 2026-03-10T10:21:54.471 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T10:21:54.471 INFO:tasks.cephadm.mon.c:Stopped mon.b 2026-03-10T10:21:54.471 INFO:tasks.cephadm.mon.c:Stopping mon.c... 2026-03-10T10:21:54.471 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mon.c 2026-03-10T10:21:54.623 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mon.c.service' 2026-03-10T10:21:54.633 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T10:21:54.633 INFO:tasks.cephadm.mon.c:Stopped mon.c 2026-03-10T10:21:54.633 INFO:tasks.cephadm.mgr.a:Stopping mgr.a... 2026-03-10T10:21:54.633 DEBUG:teuthology.orchestra.run.vm00:> sudo systemctl stop ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mgr.a 2026-03-10T10:21:54.809 DEBUG:teuthology.orchestra.run.vm00:> sudo pkill -f 'journalctl -f -n 0 -u ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mgr.a.service' 2026-03-10T10:21:54.819 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T10:21:54.819 INFO:tasks.cephadm.mgr.a:Stopped mgr.a 2026-03-10T10:21:54.820 INFO:tasks.cephadm.mgr.b:Stopping mgr.b... 2026-03-10T10:21:54.820 DEBUG:teuthology.orchestra.run.vm03:> sudo systemctl stop ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mgr.b 2026-03-10T10:21:54.945 DEBUG:teuthology.orchestra.run.vm03:> sudo pkill -f 'journalctl -f -n 0 -u ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@mgr.b.service' 2026-03-10T10:21:54.955 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T10:21:54.955 INFO:tasks.cephadm.mgr.b:Stopped mgr.b 2026-03-10T10:21:54.955 INFO:tasks.cephadm.osd.0:Stopping osd.0... 2026-03-10T10:21:54.955 DEBUG:teuthology.orchestra.run.vm00:> sudo systemctl stop ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@osd.0 2026-03-10T10:21:55.415 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 10 10:21:54 vm00 systemd[1]: Stopping Ceph osd.0 for 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad... 2026-03-10T10:21:55.415 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 10 10:21:55 vm00 bash[29585]: debug 2026-03-10T10:21:55.000+0000 7fc951496640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug (PID: 1) UID: 0 2026-03-10T10:21:55.415 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 10 10:21:55 vm00 bash[29585]: debug 2026-03-10T10:21:55.000+0000 7fc951496640 -1 osd.0 22 *** Got signal Terminated *** 2026-03-10T10:21:55.415 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 10 10:21:55 vm00 bash[29585]: debug 2026-03-10T10:21:55.000+0000 7fc951496640 -1 osd.0 22 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T10:22:00.404 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 10 10:22:00 vm00 bash[36892]: ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad-osd-0 2026-03-10T10:22:00.455 DEBUG:teuthology.orchestra.run.vm00:> sudo pkill -f 'journalctl -f -n 0 -u ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@osd.0.service' 2026-03-10T10:22:00.477 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T10:22:00.477 INFO:tasks.cephadm.osd.0:Stopped osd.0 2026-03-10T10:22:00.477 INFO:tasks.cephadm.osd.1:Stopping osd.1... 2026-03-10T10:22:00.478 DEBUG:teuthology.orchestra.run.vm03:> sudo systemctl stop ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@osd.1 2026-03-10T10:22:00.830 INFO:journalctl@ceph.osd.1.vm03.stdout:Mar 10 10:22:00 vm03 systemd[1]: Stopping Ceph osd.1 for 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad... 2026-03-10T10:22:00.831 INFO:journalctl@ceph.osd.1.vm03.stdout:Mar 10 10:22:00 vm03 bash[25520]: debug 2026-03-10T10:22:00.522+0000 7f6c770cf640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug (PID: 1) UID: 0 2026-03-10T10:22:00.831 INFO:journalctl@ceph.osd.1.vm03.stdout:Mar 10 10:22:00 vm03 bash[25520]: debug 2026-03-10T10:22:00.522+0000 7f6c770cf640 -1 osd.1 22 *** Got signal Terminated *** 2026-03-10T10:22:00.831 INFO:journalctl@ceph.osd.1.vm03.stdout:Mar 10 10:22:00 vm03 bash[25520]: debug 2026-03-10T10:22:00.522+0000 7f6c770cf640 -1 osd.1 22 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T10:22:05.830 INFO:journalctl@ceph.osd.1.vm03.stdout:Mar 10 10:22:05 vm03 bash[30347]: ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad-osd-1 2026-03-10T10:22:05.920 DEBUG:teuthology.orchestra.run.vm03:> sudo pkill -f 'journalctl -f -n 0 -u ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@osd.1.service' 2026-03-10T10:22:05.943 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T10:22:05.943 INFO:tasks.cephadm.osd.1:Stopped osd.1 2026-03-10T10:22:05.943 INFO:tasks.cephadm.osd.2:Stopping osd.2... 2026-03-10T10:22:05.943 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@osd.2 2026-03-10T10:22:06.258 INFO:journalctl@ceph.osd.2.vm09.stdout:Mar 10 10:22:05 vm09 systemd[1]: Stopping Ceph osd.2 for 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad... 2026-03-10T10:22:06.258 INFO:journalctl@ceph.osd.2.vm09.stdout:Mar 10 10:22:05 vm09 bash[25744]: debug 2026-03-10T10:22:05.985+0000 7ff93eed8640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug (PID: 1) UID: 0 2026-03-10T10:22:06.258 INFO:journalctl@ceph.osd.2.vm09.stdout:Mar 10 10:22:05 vm09 bash[25744]: debug 2026-03-10T10:22:05.985+0000 7ff93eed8640 -1 osd.2 22 *** Got signal Terminated *** 2026-03-10T10:22:06.258 INFO:journalctl@ceph.osd.2.vm09.stdout:Mar 10 10:22:05 vm09 bash[25744]: debug 2026-03-10T10:22:05.985+0000 7ff93eed8640 -1 osd.2 22 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T10:22:11.337 INFO:journalctl@ceph.osd.2.vm09.stdout:Mar 10 10:22:11 vm09 bash[30090]: ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad-osd-2 2026-03-10T10:22:11.387 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad@osd.2.service' 2026-03-10T10:22:11.411 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T10:22:11.411 INFO:tasks.cephadm.osd.2:Stopped osd.2 2026-03-10T10:22:11.412 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad --force --keep-logs 2026-03-10T10:22:11.506 INFO:teuthology.orchestra.run.vm00.stdout:Deleting cluster with fsid: 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:22:13.780 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad --force --keep-logs 2026-03-10T10:22:13.871 INFO:teuthology.orchestra.run.vm03.stdout:Deleting cluster with fsid: 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:22:16.114 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad --force --keep-logs 2026-03-10T10:22:16.211 INFO:teuthology.orchestra.run.vm09.stdout:Deleting cluster with fsid: 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:22:18.282 DEBUG:teuthology.orchestra.run.vm00:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:22:18.289 INFO:teuthology.orchestra.run.vm00.stderr:rm: cannot remove '/etc/ceph/ceph.client.admin.keyring': Is a directory 2026-03-10T10:22:18.289 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:22:18.289 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:22:18.297 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:22:18.304 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-10T10:22:18.304 DEBUG:teuthology.misc:Transferring archived files from vm00:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/crash to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/996/remote/vm00/crash 2026-03-10T10:22:18.304 DEBUG:teuthology.orchestra.run.vm00:> sudo tar c -f - -C /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/crash -- . 2026-03-10T10:22:18.337 INFO:teuthology.orchestra.run.vm00.stderr:tar: /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/crash: Cannot open: No such file or directory 2026-03-10T10:22:18.337 INFO:teuthology.orchestra.run.vm00.stderr:tar: Error is not recoverable: exiting now 2026-03-10T10:22:18.338 DEBUG:teuthology.misc:Transferring archived files from vm03:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/crash to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/996/remote/vm03/crash 2026-03-10T10:22:18.338 DEBUG:teuthology.orchestra.run.vm03:> sudo tar c -f - -C /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/crash -- . 2026-03-10T10:22:18.345 INFO:teuthology.orchestra.run.vm03.stderr:tar: /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/crash: Cannot open: No such file or directory 2026-03-10T10:22:18.345 INFO:teuthology.orchestra.run.vm03.stderr:tar: Error is not recoverable: exiting now 2026-03-10T10:22:18.346 DEBUG:teuthology.misc:Transferring archived files from vm09:/var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/crash to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/996/remote/vm09/crash 2026-03-10T10:22:18.346 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/crash -- . 2026-03-10T10:22:18.353 INFO:teuthology.orchestra.run.vm09.stderr:tar: /var/lib/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/crash: Cannot open: No such file or directory 2026-03-10T10:22:18.353 INFO:teuthology.orchestra.run.vm09.stderr:tar: Error is not recoverable: exiting now 2026-03-10T10:22:18.353 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-10T10:22:18.353 DEBUG:teuthology.orchestra.run.vm00:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v MON_DOWN | egrep -v POOL_APP_NOT_ENABLED | egrep -v 'mon down' | egrep -v 'mons down' | egrep -v 'out of quorum' | egrep -v CEPHADM_FAILED_DAEMON | head -n 1 2026-03-10T10:22:18.389 INFO:tasks.cephadm:Compressing logs... 2026-03-10T10:22:18.390 DEBUG:teuthology.orchestra.run.vm00:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T10:22:18.432 DEBUG:teuthology.orchestra.run.vm03:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T10:22:18.433 DEBUG:teuthology.orchestra.run.vm09:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T10:22:18.437 INFO:teuthology.orchestra.run.vm00.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-10T10:22:18.438 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-10T10:22:18.439 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-mgr.a.log 2026-03-10T10:22:18.439 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.log 2026-03-10T10:22:18.441 INFO:teuthology.orchestra.run.vm03.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-10T10:22:18.441 INFO:teuthology.orchestra.run.vm03.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-10T10:22:18.441 INFO:teuthology.orchestra.run.vm09.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-10T10:22:18.441 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-mgr.a.log: 87.4% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-10T10:22:18.441 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-mon.a.log 2026-03-10T10:22:18.441 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.log 2026-03-10T10:22:18.442 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.log: 87.5% -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.log.gz 2026-03-10T10:22:18.442 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.audit.log 2026-03-10T10:22:18.442 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-mon.b.log 2026-03-10T10:22:18.442 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-10T10:22:18.442 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.log 2026-03-10T10:22:18.443 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.log: 85.7% -- replaced with /var/log/ceph/cephadm.log.gz 87.6% 2026-03-10T10:22:18.443 INFO:teuthology.orchestra.run.vm03.stderr: -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.log.gz 2026-03-10T10:22:18.443 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-mon.c.log 2026-03-10T10:22:18.443 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-osd.1.log 2026-03-10T10:22:18.443 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-mon.b.log: gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-mgr.b.log 2026-03-10T10:22:18.444 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.log: 87.6% 84.8% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-10T10:22:18.444 INFO:teuthology.orchestra.run.vm09.stderr: -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.log.gz 2026-03-10T10:22:18.444 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-osd.2.log 2026-03-10T10:22:18.444 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-mon.c.log: gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.audit.log 2026-03-10T10:22:18.448 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-mon.a.log: gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-volume.log 2026-03-10T10:22:18.449 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.audit.log: 90.5% -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.audit.log.gz 2026-03-10T10:22:18.451 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-osd.1.log: gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.audit.log 2026-03-10T10:22:18.452 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.cephadm.log 2026-03-10T10:22:18.454 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-mgr.b.log: 91.5% -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-mgr.b.log.gz 2026-03-10T10:22:18.454 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-volume.log 2026-03-10T10:22:18.454 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-osd.2.log: gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-volume.log 2026-03-10T10:22:18.455 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.audit.log: 90.8% -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.audit.log.gz 2026-03-10T10:22:18.456 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.cephadm.log 2026-03-10T10:22:18.459 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.audit.log: 90.8% -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.audit.log.gz 2026-03-10T10:22:18.459 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.cephadm.log 2026-03-10T10:22:18.463 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-volume.log: 93.5% -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-osd.2.log.gz 2026-03-10T10:22:18.463 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.cephadm.log: 82.8% -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.cephadm.log.gz 2026-03-10T10:22:18.464 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-volume.log: gzip -5 --verbose -- /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-osd.0.log 2026-03-10T10:22:18.465 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.cephadm.log: 82.6% 96.1% -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.cephadm.log.gz 2026-03-10T10:22:18.465 INFO:teuthology.orchestra.run.vm00.stderr: -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-volume.log.gz 2026-03-10T10:22:18.466 INFO:teuthology.orchestra.run.vm09.stderr: 96.0% -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-volume.log.gz 2026-03-10T10:22:18.469 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-volume.log: 93.6% -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-osd.1.log.gz 2026-03-10T10:22:18.470 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.cephadm.log: 82.8% -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph.cephadm.log.gz 2026-03-10T10:22:18.470 INFO:teuthology.orchestra.run.vm03.stderr: 96.0% -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-volume.log.gz 2026-03-10T10:22:18.491 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-osd.0.log: 89.2% -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-mgr.a.log.gz 2026-03-10T10:22:18.501 INFO:teuthology.orchestra.run.vm03.stderr: 93.2% -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-mon.b.log.gz 2026-03-10T10:22:18.501 INFO:teuthology.orchestra.run.vm00.stderr: 93.5% -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-osd.0.log.gz 2026-03-10T10:22:18.502 INFO:teuthology.orchestra.run.vm03.stderr: 2026-03-10T10:22:18.503 INFO:teuthology.orchestra.run.vm03.stderr:real 0m0.067s 2026-03-10T10:22:18.503 INFO:teuthology.orchestra.run.vm03.stderr:user 0m0.087s 2026-03-10T10:22:18.503 INFO:teuthology.orchestra.run.vm03.stderr:sys 0m0.010s 2026-03-10T10:22:18.504 INFO:teuthology.orchestra.run.vm09.stderr: 92.7% -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-mon.c.log.gz 2026-03-10T10:22:18.505 INFO:teuthology.orchestra.run.vm09.stderr: 2026-03-10T10:22:18.505 INFO:teuthology.orchestra.run.vm09.stderr:real 0m0.069s 2026-03-10T10:22:18.505 INFO:teuthology.orchestra.run.vm09.stderr:user 0m0.076s 2026-03-10T10:22:18.505 INFO:teuthology.orchestra.run.vm09.stderr:sys 0m0.018s 2026-03-10T10:22:18.617 INFO:teuthology.orchestra.run.vm00.stderr: 91.6% -- replaced with /var/log/ceph/9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad/ceph-mon.a.log.gz 2026-03-10T10:22:18.619 INFO:teuthology.orchestra.run.vm00.stderr: 2026-03-10T10:22:18.619 INFO:teuthology.orchestra.run.vm00.stderr:real 0m0.185s 2026-03-10T10:22:18.619 INFO:teuthology.orchestra.run.vm00.stderr:user 0m0.237s 2026-03-10T10:22:18.619 INFO:teuthology.orchestra.run.vm00.stderr:sys 0m0.013s 2026-03-10T10:22:18.619 INFO:tasks.cephadm:Archiving logs... 2026-03-10T10:22:18.619 DEBUG:teuthology.misc:Transferring archived files from vm00:/var/log/ceph to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/996/remote/vm00/log 2026-03-10T10:22:18.619 DEBUG:teuthology.orchestra.run.vm00:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-10T10:22:18.682 DEBUG:teuthology.misc:Transferring archived files from vm03:/var/log/ceph to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/996/remote/vm03/log 2026-03-10T10:22:18.682 DEBUG:teuthology.orchestra.run.vm03:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-10T10:22:18.696 DEBUG:teuthology.misc:Transferring archived files from vm09:/var/log/ceph to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/996/remote/vm09/log 2026-03-10T10:22:18.696 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-10T10:22:18.708 INFO:tasks.cephadm:Removing cluster... 2026-03-10T10:22:18.708 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad --force 2026-03-10T10:22:18.809 INFO:teuthology.orchestra.run.vm00.stdout:Deleting cluster with fsid: 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:22:20.015 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad --force 2026-03-10T10:22:20.102 INFO:teuthology.orchestra.run.vm03.stdout:Deleting cluster with fsid: 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:22:21.318 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad --force 2026-03-10T10:22:21.409 INFO:teuthology.orchestra.run.vm09.stdout:Deleting cluster with fsid: 9cae97e0-1c69-11f1-b5b6-cf0a2e9653ad 2026-03-10T10:22:22.612 INFO:tasks.cephadm:Removing cephadm ... 2026-03-10T10:22:22.612 DEBUG:teuthology.orchestra.run.vm00:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-10T10:22:22.615 DEBUG:teuthology.orchestra.run.vm03:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-10T10:22:22.618 DEBUG:teuthology.orchestra.run.vm09:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-10T10:22:22.620 INFO:tasks.cephadm:Teardown complete 2026-03-10T10:22:22.621 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-03-10T10:22:22.623 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-03-10T10:22:22.623 DEBUG:teuthology.orchestra.run.vm00:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-10T10:22:22.657 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-10T10:22:22.663 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-10T10:22:22.676 INFO:teuthology.task.install.deb:Removing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on Debian system. 2026-03-10T10:22:22.677 DEBUG:teuthology.orchestra.run.vm00:> for d in ceph cephadm ceph-mds ceph-mgr ceph-common ceph-fuse ceph-test ceph-volume radosgw python3-rados python3-rgw python3-cephfs python3-rbd libcephfs2 libcephfs-dev librados2 librbd1 rbd-fuse ; do sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" purge $d || true ; done 2026-03-10T10:22:22.682 INFO:teuthology.task.install.deb:Removing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on Debian system. 2026-03-10T10:22:22.682 DEBUG:teuthology.orchestra.run.vm03:> for d in ceph cephadm ceph-mds ceph-mgr ceph-common ceph-fuse ceph-test ceph-volume radosgw python3-rados python3-rgw python3-cephfs python3-rbd libcephfs2 libcephfs-dev librados2 librbd1 rbd-fuse ; do sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" purge $d || true ; done 2026-03-10T10:22:22.687 INFO:teuthology.task.install.deb:Removing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on Debian system. 2026-03-10T10:22:22.687 DEBUG:teuthology.orchestra.run.vm09:> for d in ceph cephadm ceph-mds ceph-mgr ceph-common ceph-fuse ceph-test ceph-volume radosgw python3-rados python3-rgw python3-cephfs python3-rbd libcephfs2 libcephfs-dev librados2 librbd1 rbd-fuse ; do sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" purge $d || true ; done 2026-03-10T10:22:22.741 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:22.744 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:22.748 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:22.937 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:22.938 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:22.941 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:22.941 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:22.942 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:22.942 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:23.124 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:23.124 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T10:22:23.124 INFO:teuthology.orchestra.run.vm03.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-10T10:22:23.124 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:23.139 INFO:teuthology.orchestra.run.vm03.stdout:The following packages will be REMOVED: 2026-03-10T10:22:23.139 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:23.139 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T10:22:23.140 INFO:teuthology.orchestra.run.vm00.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-10T10:22:23.140 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:23.140 INFO:teuthology.orchestra.run.vm03.stdout: ceph* 2026-03-10T10:22:23.156 INFO:teuthology.orchestra.run.vm00.stdout:The following packages will be REMOVED: 2026-03-10T10:22:23.157 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:23.157 INFO:teuthology.orchestra.run.vm00.stdout: ceph* 2026-03-10T10:22:23.157 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T10:22:23.158 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-10T10:22:23.158 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:23.174 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-10T10:22:23.175 INFO:teuthology.orchestra.run.vm09.stdout: ceph* 2026-03-10T10:22:23.326 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T10:22:23.326 INFO:teuthology.orchestra.run.vm03.stdout:After this operation, 47.1 kB disk space will be freed. 2026-03-10T10:22:23.332 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T10:22:23.332 INFO:teuthology.orchestra.run.vm00.stdout:After this operation, 47.1 kB disk space will be freed. 2026-03-10T10:22:23.348 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T10:22:23.348 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 47.1 kB disk space will be freed. 2026-03-10T10:22:23.366 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118605 files and directories currently installed.) 2026-03-10T10:22:23.368 INFO:teuthology.orchestra.run.vm03.stdout:Removing ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:23.369 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118605 files and directories currently installed.) 2026-03-10T10:22:23.371 INFO:teuthology.orchestra.run.vm00.stdout:Removing ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:23.389 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118605 files and directories currently installed.) 2026-03-10T10:22:23.392 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:24.552 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:24.572 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:24.573 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:24.587 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:24.609 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:24.611 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:24.791 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:24.792 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:24.813 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:24.814 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:24.823 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:24.823 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:24.991 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:24.992 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T10:22:24.992 INFO:teuthology.orchestra.run.vm03.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-10T10:22:24.992 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:25.006 INFO:teuthology.orchestra.run.vm03.stdout:The following packages will be REMOVED: 2026-03-10T10:22:25.007 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-cephadm* cephadm* 2026-03-10T10:22:25.034 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:25.034 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T10:22:25.034 INFO:teuthology.orchestra.run.vm00.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-10T10:22:25.034 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:25.046 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:25.051 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T10:22:25.051 INFO:teuthology.orchestra.run.vm00.stdout:The following packages will be REMOVED: 2026-03-10T10:22:25.051 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-10T10:22:25.052 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:25.052 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-cephadm* cephadm* 2026-03-10T10:22:25.067 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-10T10:22:25.068 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-cephadm* cephadm* 2026-03-10T10:22:25.187 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 2 to remove and 10 not upgraded. 2026-03-10T10:22:25.187 INFO:teuthology.orchestra.run.vm03.stdout:After this operation, 1775 kB disk space will be freed. 2026-03-10T10:22:25.229 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118603 files and directories currently installed.) 2026-03-10T10:22:25.230 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 2 to remove and 10 not upgraded. 2026-03-10T10:22:25.230 INFO:teuthology.orchestra.run.vm00.stdout:After this operation, 1775 kB disk space will be freed. 2026-03-10T10:22:25.231 INFO:teuthology.orchestra.run.vm03.stdout:Removing ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:25.278 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 2 to remove and 10 not upgraded. 2026-03-10T10:22:25.278 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 1775 kB disk space will be freed. 2026-03-10T10:22:25.305 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118603 files and directories currently installed.) 2026-03-10T10:22:25.317 INFO:teuthology.orchestra.run.vm00.stdout:Removing ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:25.409 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118603 files and directories currently installed.) 2026-03-10T10:22:25.412 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-mgr-cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:25.756 INFO:teuthology.orchestra.run.vm03.stdout:Removing cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:25.758 INFO:teuthology.orchestra.run.vm00.stdout:Removing cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:25.760 INFO:teuthology.orchestra.run.vm09.stdout:Removing cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:25.786 INFO:teuthology.orchestra.run.vm03.stdout:Looking for files to backup/remove ... 2026-03-10T10:22:25.787 INFO:teuthology.orchestra.run.vm03.stdout:Not backing up/removing `/var/lib/cephadm', it matches ^/var/.*. 2026-03-10T10:22:25.789 INFO:teuthology.orchestra.run.vm03.stdout:Removing user `cephadm' ... 2026-03-10T10:22:25.789 INFO:teuthology.orchestra.run.vm03.stdout:Warning: group `nogroup' has no more members. 2026-03-10T10:22:25.792 INFO:teuthology.orchestra.run.vm09.stdout:Looking for files to backup/remove ... 2026-03-10T10:22:25.794 INFO:teuthology.orchestra.run.vm09.stdout:Not backing up/removing `/var/lib/cephadm', it matches ^/var/.*. 2026-03-10T10:22:25.795 INFO:teuthology.orchestra.run.vm00.stdout:Looking for files to backup/remove ... 2026-03-10T10:22:25.796 INFO:teuthology.orchestra.run.vm09.stdout:Removing user `cephadm' ... 2026-03-10T10:22:25.796 INFO:teuthology.orchestra.run.vm09.stdout:Warning: group `nogroup' has no more members. 2026-03-10T10:22:25.796 INFO:teuthology.orchestra.run.vm00.stdout:Not backing up/removing `/var/lib/cephadm', it matches ^/var/.*. 2026-03-10T10:22:25.798 INFO:teuthology.orchestra.run.vm03.stdout:Done. 2026-03-10T10:22:25.799 INFO:teuthology.orchestra.run.vm00.stdout:Removing user `cephadm' ... 2026-03-10T10:22:25.799 INFO:teuthology.orchestra.run.vm00.stdout:Warning: group `nogroup' has no more members. 2026-03-10T10:22:25.807 INFO:teuthology.orchestra.run.vm09.stdout:Done. 2026-03-10T10:22:25.811 INFO:teuthology.orchestra.run.vm00.stdout:Done. 2026-03-10T10:22:25.819 INFO:teuthology.orchestra.run.vm03.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:25.830 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:25.834 INFO:teuthology.orchestra.run.vm00.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:25.930 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118529 files and directories currently installed.) 2026-03-10T10:22:25.931 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118529 files and directories currently installed.) 2026-03-10T10:22:25.932 INFO:teuthology.orchestra.run.vm00.stdout:Purging configuration files for cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:25.933 INFO:teuthology.orchestra.run.vm03.stdout:Purging configuration files for cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:25.947 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118529 files and directories currently installed.) 2026-03-10T10:22:25.949 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for cephadm (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:27.050 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:27.074 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:27.081 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:27.083 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:27.107 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:27.119 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:27.281 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:27.281 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:27.299 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:27.300 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:27.325 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:27.326 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:27.510 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:27.510 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T10:22:27.510 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-10T10:22:27.510 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:27.520 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:27.520 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T10:22:27.520 INFO:teuthology.orchestra.run.vm00.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-10T10:22:27.521 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:27.524 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-10T10:22:27.524 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mds* 2026-03-10T10:22:27.534 INFO:teuthology.orchestra.run.vm00.stdout:The following packages will be REMOVED: 2026-03-10T10:22:27.535 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mds* 2026-03-10T10:22:27.571 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:27.571 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mon kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-10T10:22:27.572 INFO:teuthology.orchestra.run.vm03.stdout: libsgutils2-2 python-asyncssh-doc python3-asyncssh sg3-utils sg3-utils-udev 2026-03-10T10:22:27.572 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:27.587 INFO:teuthology.orchestra.run.vm03.stdout:The following packages will be REMOVED: 2026-03-10T10:22:27.588 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mds* 2026-03-10T10:22:27.699 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T10:22:27.699 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 7437 kB disk space will be freed. 2026-03-10T10:22:27.730 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T10:22:27.731 INFO:teuthology.orchestra.run.vm00.stdout:After this operation, 7437 kB disk space will be freed. 2026-03-10T10:22:27.734 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118529 files and directories currently installed.) 2026-03-10T10:22:27.736 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:27.763 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118529 files and directories currently installed.) 2026-03-10T10:22:27.765 INFO:teuthology.orchestra.run.vm00.stdout:Removing ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:27.777 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T10:22:27.778 INFO:teuthology.orchestra.run.vm03.stdout:After this operation, 7437 kB disk space will be freed. 2026-03-10T10:22:27.817 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118529 files and directories currently installed.) 2026-03-10T10:22:27.819 INFO:teuthology.orchestra.run.vm03.stdout:Removing ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:28.178 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:28.182 INFO:teuthology.orchestra.run.vm00.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:28.285 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118521 files and directories currently installed.) 2026-03-10T10:22:28.287 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118521 files and directories currently installed.) 2026-03-10T10:22:28.288 INFO:teuthology.orchestra.run.vm00.stdout:Purging configuration files for ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:28.290 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:28.312 INFO:teuthology.orchestra.run.vm03.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:28.413 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118521 files and directories currently installed.) 2026-03-10T10:22:28.415 INFO:teuthology.orchestra.run.vm03.stdout:Purging configuration files for ceph-mds (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:29.867 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:29.882 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:29.905 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:29.920 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:29.997 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:30.031 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:30.107 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:30.108 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:30.116 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:30.116 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:30.233 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:30.234 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:30.287 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:30.288 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core ceph-mon kpartx libboost-iostreams1.74.0 2026-03-10T10:22:30.289 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libpmemobj1 libsgutils2-2 python-asyncssh-doc 2026-03-10T10:22:30.289 INFO:teuthology.orchestra.run.vm09.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools python3-cheroot 2026-03-10T10:22:30.289 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:30.289 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:30.289 INFO:teuthology.orchestra.run.vm09.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:30.289 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:30.289 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan python3-portend python3-psutil python3-pyinotify 2026-03-10T10:22:30.289 INFO:teuthology.orchestra.run.vm09.stdout: python3-repoze.lru python3-requests-oauthlib python3-routes python3-rsa 2026-03-10T10:22:30.289 INFO:teuthology.orchestra.run.vm09.stdout: python3-simplegeneric python3-simplejson python3-singledispatch 2026-03-10T10:22:30.289 INFO:teuthology.orchestra.run.vm09.stdout: python3-sklearn python3-sklearn-lib python3-tempita python3-tempora 2026-03-10T10:22:30.289 INFO:teuthology.orchestra.run.vm09.stdout: python3-threadpoolctl python3-waitress python3-webob python3-websocket 2026-03-10T10:22:30.289 INFO:teuthology.orchestra.run.vm09.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:30.289 INFO:teuthology.orchestra.run.vm09.stdout: sg3-utils-udev 2026-03-10T10:22:30.289 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:30.292 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:30.293 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core ceph-mon kpartx libboost-iostreams1.74.0 2026-03-10T10:22:30.294 INFO:teuthology.orchestra.run.vm00.stdout: libboost-thread1.74.0 libpmemobj1 libsgutils2-2 python-asyncssh-doc 2026-03-10T10:22:30.294 INFO:teuthology.orchestra.run.vm00.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools python3-cheroot 2026-03-10T10:22:30.294 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:30.294 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:30.294 INFO:teuthology.orchestra.run.vm00.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:30.294 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:30.294 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan python3-portend python3-psutil python3-pyinotify 2026-03-10T10:22:30.294 INFO:teuthology.orchestra.run.vm00.stdout: python3-repoze.lru python3-requests-oauthlib python3-routes python3-rsa 2026-03-10T10:22:30.294 INFO:teuthology.orchestra.run.vm00.stdout: python3-simplegeneric python3-simplejson python3-singledispatch 2026-03-10T10:22:30.294 INFO:teuthology.orchestra.run.vm00.stdout: python3-sklearn python3-sklearn-lib python3-tempita python3-tempora 2026-03-10T10:22:30.294 INFO:teuthology.orchestra.run.vm00.stdout: python3-threadpoolctl python3-waitress python3-webob python3-websocket 2026-03-10T10:22:30.294 INFO:teuthology.orchestra.run.vm00.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:30.294 INFO:teuthology.orchestra.run.vm00.stdout: sg3-utils-udev 2026-03-10T10:22:30.294 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:30.303 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-10T10:22:30.303 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr* ceph-mgr-dashboard* ceph-mgr-diskprediction-local* 2026-03-10T10:22:30.304 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-k8sevents* 2026-03-10T10:22:30.309 INFO:teuthology.orchestra.run.vm00.stdout:The following packages will be REMOVED: 2026-03-10T10:22:30.309 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr* ceph-mgr-dashboard* ceph-mgr-diskprediction-local* 2026-03-10T10:22:30.310 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-k8sevents* 2026-03-10T10:22:30.385 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:30.385 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core ceph-mon kpartx libboost-iostreams1.74.0 2026-03-10T10:22:30.386 INFO:teuthology.orchestra.run.vm03.stdout: libboost-thread1.74.0 libpmemobj1 libsgutils2-2 python-asyncssh-doc 2026-03-10T10:22:30.386 INFO:teuthology.orchestra.run.vm03.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools python3-cheroot 2026-03-10T10:22:30.386 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:30.386 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:30.386 INFO:teuthology.orchestra.run.vm03.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:30.386 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:30.386 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan python3-portend python3-psutil python3-pyinotify 2026-03-10T10:22:30.386 INFO:teuthology.orchestra.run.vm03.stdout: python3-repoze.lru python3-requests-oauthlib python3-routes python3-rsa 2026-03-10T10:22:30.386 INFO:teuthology.orchestra.run.vm03.stdout: python3-simplegeneric python3-simplejson python3-singledispatch 2026-03-10T10:22:30.386 INFO:teuthology.orchestra.run.vm03.stdout: python3-sklearn python3-sklearn-lib python3-tempita python3-tempora 2026-03-10T10:22:30.386 INFO:teuthology.orchestra.run.vm03.stdout: python3-threadpoolctl python3-waitress python3-webob python3-websocket 2026-03-10T10:22:30.386 INFO:teuthology.orchestra.run.vm03.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:30.386 INFO:teuthology.orchestra.run.vm03.stdout: sg3-utils-udev 2026-03-10T10:22:30.386 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:30.397 INFO:teuthology.orchestra.run.vm03.stdout:The following packages will be REMOVED: 2026-03-10T10:22:30.397 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr* ceph-mgr-dashboard* ceph-mgr-diskprediction-local* 2026-03-10T10:22:30.398 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-k8sevents* 2026-03-10T10:22:30.507 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 4 to remove and 10 not upgraded. 2026-03-10T10:22:30.507 INFO:teuthology.orchestra.run.vm00.stdout:After this operation, 165 MB disk space will be freed. 2026-03-10T10:22:30.539 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 4 to remove and 10 not upgraded. 2026-03-10T10:22:30.539 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 165 MB disk space will be freed. 2026-03-10T10:22:30.591 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118521 files and directories currently installed.) 2026-03-10T10:22:30.593 INFO:teuthology.orchestra.run.vm00.stdout:Removing ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:30.633 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 4 to remove and 10 not upgraded. 2026-03-10T10:22:30.633 INFO:teuthology.orchestra.run.vm03.stdout:After this operation, 165 MB disk space will be freed. 2026-03-10T10:22:30.641 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118521 files and directories currently installed.) 2026-03-10T10:22:30.643 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:30.712 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 118521 files and directories currently installed.) 2026-03-10T10:22:30.714 INFO:teuthology.orchestra.run.vm03.stdout:Removing ceph-mgr-k8sevents (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:30.731 INFO:teuthology.orchestra.run.vm00.stdout:Removing ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:30.734 INFO:teuthology.orchestra.run.vm03.stdout:Removing ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:30.740 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-mgr-diskprediction-local (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:30.758 INFO:teuthology.orchestra.run.vm00.stdout:Removing ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:30.762 INFO:teuthology.orchestra.run.vm03.stdout:Removing ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:30.771 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-mgr-dashboard (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:30.964 INFO:teuthology.orchestra.run.vm00.stdout:Removing ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:30.968 INFO:teuthology.orchestra.run.vm03.stdout:Removing ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:31.126 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:31.777 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117937 files and directories currently installed.) 2026-03-10T10:22:31.780 INFO:teuthology.orchestra.run.vm03.stdout:Purging configuration files for ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:31.783 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117937 files and directories currently installed.) 2026-03-10T10:22:31.786 INFO:teuthology.orchestra.run.vm00.stdout:Purging configuration files for ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:31.786 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117937 files and directories currently installed.) 2026-03-10T10:22:31.788 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for ceph-mgr (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:33.229 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:33.261 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:33.288 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:33.295 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:33.320 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:33.329 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:33.456 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:33.457 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:33.530 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:33.531 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:33.541 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:33.541 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:33.592 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:33.592 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:33.592 INFO:teuthology.orchestra.run.vm03.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T10:22:33.592 INFO:teuthology.orchestra.run.vm03.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T10:22:33.592 INFO:teuthology.orchestra.run.vm03.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:22:33.592 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T10:22:33.592 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:22:33.592 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:22:33.592 INFO:teuthology.orchestra.run.vm03.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:22:33.592 INFO:teuthology.orchestra.run.vm03.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T10:22:33.592 INFO:teuthology.orchestra.run.vm03.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T10:22:33.592 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:22:33.592 INFO:teuthology.orchestra.run.vm03.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:22:33.592 INFO:teuthology.orchestra.run.vm03.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:22:33.592 INFO:teuthology.orchestra.run.vm03.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T10:22:33.592 INFO:teuthology.orchestra.run.vm03.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:33.592 INFO:teuthology.orchestra.run.vm03.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T10:22:33.593 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:33.604 INFO:teuthology.orchestra.run.vm03.stdout:The following packages will be REMOVED: 2026-03-10T10:22:33.605 INFO:teuthology.orchestra.run.vm03.stdout: ceph-base* ceph-common* ceph-mon* ceph-osd* ceph-test* ceph-volume* radosgw* 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T10:22:33.680 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:33.690 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-10T10:22:33.690 INFO:teuthology.orchestra.run.vm09.stdout: ceph-base* ceph-common* ceph-mon* ceph-osd* ceph-test* ceph-volume* radosgw* 2026-03-10T10:22:33.704 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:33.704 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:33.704 INFO:teuthology.orchestra.run.vm00.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T10:22:33.705 INFO:teuthology.orchestra.run.vm00.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T10:22:33.705 INFO:teuthology.orchestra.run.vm00.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:22:33.705 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T10:22:33.705 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:22:33.705 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:22:33.705 INFO:teuthology.orchestra.run.vm00.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:22:33.705 INFO:teuthology.orchestra.run.vm00.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T10:22:33.705 INFO:teuthology.orchestra.run.vm00.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T10:22:33.705 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:22:33.705 INFO:teuthology.orchestra.run.vm00.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:22:33.705 INFO:teuthology.orchestra.run.vm00.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:22:33.705 INFO:teuthology.orchestra.run.vm00.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T10:22:33.705 INFO:teuthology.orchestra.run.vm00.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:33.705 INFO:teuthology.orchestra.run.vm00.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T10:22:33.705 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:33.713 INFO:teuthology.orchestra.run.vm00.stdout:The following packages will be REMOVED: 2026-03-10T10:22:33.714 INFO:teuthology.orchestra.run.vm00.stdout: ceph-base* ceph-common* ceph-mon* ceph-osd* ceph-test* ceph-volume* radosgw* 2026-03-10T10:22:33.910 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 7 to remove and 10 not upgraded. 2026-03-10T10:22:33.913 INFO:teuthology.orchestra.run.vm03.stdout:After this operation, 472 MB disk space will be freed. 2026-03-10T10:22:33.953 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 7 to remove and 10 not upgraded. 2026-03-10T10:22:33.953 INFO:teuthology.orchestra.run.vm00.stdout:After this operation, 472 MB disk space will be freed. 2026-03-10T10:22:33.956 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 7 to remove and 10 not upgraded. 2026-03-10T10:22:33.956 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 472 MB disk space will be freed. 2026-03-10T10:22:33.976 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117937 files and directories currently installed.) 2026-03-10T10:22:33.978 INFO:teuthology.orchestra.run.vm03.stdout:Removing ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:33.993 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117937 files and directories currently installed.) 2026-03-10T10:22:33.994 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:33.996 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117937 files and directories currently installed.) 2026-03-10T10:22:33.999 INFO:teuthology.orchestra.run.vm00.stdout:Removing ceph-volume (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:34.042 INFO:teuthology.orchestra.run.vm03.stdout:Removing ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:34.057 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:34.062 INFO:teuthology.orchestra.run.vm00.stdout:Removing ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:34.424 INFO:teuthology.orchestra.run.vm03.stdout:Removing ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:34.457 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:34.480 INFO:teuthology.orchestra.run.vm00.stdout:Removing ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:34.827 INFO:teuthology.orchestra.run.vm03.stdout:Removing ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:34.865 INFO:teuthology.orchestra.run.vm00.stdout:Removing ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:34.871 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:35.214 INFO:teuthology.orchestra.run.vm00.stdout:Removing radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:35.263 INFO:teuthology.orchestra.run.vm03.stdout:Removing radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:35.323 INFO:teuthology.orchestra.run.vm09.stdout:Removing radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:35.629 INFO:teuthology.orchestra.run.vm00.stdout:Removing ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:35.665 INFO:teuthology.orchestra.run.vm00.stdout:Removing ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:35.733 INFO:teuthology.orchestra.run.vm03.stdout:Removing ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:35.742 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-test (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:35.774 INFO:teuthology.orchestra.run.vm03.stdout:Removing ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:35.781 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:36.099 INFO:teuthology.orchestra.run.vm00.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:36.132 INFO:teuthology.orchestra.run.vm00.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T10:22:36.203 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117455 files and directories currently installed.) 2026-03-10T10:22:36.205 INFO:teuthology.orchestra.run.vm00.stdout:Purging configuration files for radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:36.226 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:36.237 INFO:teuthology.orchestra.run.vm03.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:36.266 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T10:22:36.272 INFO:teuthology.orchestra.run.vm03.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T10:22:36.338 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117455 files and directories currently installed.) 2026-03-10T10:22:36.340 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:36.345 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117455 files and directories currently installed.) 2026-03-10T10:22:36.347 INFO:teuthology.orchestra.run.vm03.stdout:Purging configuration files for radosgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:36.785 INFO:teuthology.orchestra.run.vm00.stdout:Purging configuration files for ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:36.916 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:36.925 INFO:teuthology.orchestra.run.vm03.stdout:Purging configuration files for ceph-mon (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:37.136 INFO:teuthology.orchestra.run.vm00.stdout:Purging configuration files for ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:37.336 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:37.362 INFO:teuthology.orchestra.run.vm03.stdout:Purging configuration files for ceph-base (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:37.534 INFO:teuthology.orchestra.run.vm00.stdout:Purging configuration files for ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:37.768 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:37.788 INFO:teuthology.orchestra.run.vm03.stdout:Purging configuration files for ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:37.966 INFO:teuthology.orchestra.run.vm00.stdout:Purging configuration files for ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:38.167 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:38.291 INFO:teuthology.orchestra.run.vm03.stdout:Purging configuration files for ceph-osd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:39.534 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:39.567 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:39.670 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:39.705 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:39.769 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:39.770 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:39.869 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:39.885 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:39.886 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:39.890 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:39.890 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:39.890 INFO:teuthology.orchestra.run.vm00.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T10:22:39.890 INFO:teuthology.orchestra.run.vm00.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T10:22:39.890 INFO:teuthology.orchestra.run.vm00.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:22:39.890 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T10:22:39.891 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:22:39.891 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:22:39.891 INFO:teuthology.orchestra.run.vm00.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:22:39.891 INFO:teuthology.orchestra.run.vm00.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T10:22:39.891 INFO:teuthology.orchestra.run.vm00.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T10:22:39.891 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:22:39.891 INFO:teuthology.orchestra.run.vm00.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:22:39.891 INFO:teuthology.orchestra.run.vm00.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:22:39.891 INFO:teuthology.orchestra.run.vm00.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T10:22:39.891 INFO:teuthology.orchestra.run.vm00.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:39.891 INFO:teuthology.orchestra.run.vm00.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T10:22:39.891 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:39.902 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:39.905 INFO:teuthology.orchestra.run.vm00.stdout:The following packages will be REMOVED: 2026-03-10T10:22:39.906 INFO:teuthology.orchestra.run.vm00.stdout: ceph-fuse* 2026-03-10T10:22:40.033 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:40.033 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:40.033 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T10:22:40.033 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T10:22:40.033 INFO:teuthology.orchestra.run.vm09.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:22:40.033 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T10:22:40.034 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:22:40.034 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:22:40.034 INFO:teuthology.orchestra.run.vm09.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:22:40.034 INFO:teuthology.orchestra.run.vm09.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T10:22:40.034 INFO:teuthology.orchestra.run.vm09.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T10:22:40.034 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:22:40.034 INFO:teuthology.orchestra.run.vm09.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:22:40.034 INFO:teuthology.orchestra.run.vm09.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:22:40.034 INFO:teuthology.orchestra.run.vm09.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T10:22:40.034 INFO:teuthology.orchestra.run.vm09.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:40.034 INFO:teuthology.orchestra.run.vm09.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T10:22:40.034 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:40.043 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-10T10:22:40.044 INFO:teuthology.orchestra.run.vm09.stdout: ceph-fuse* 2026-03-10T10:22:40.074 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T10:22:40.075 INFO:teuthology.orchestra.run.vm00.stdout:After this operation, 3673 kB disk space will be freed. 2026-03-10T10:22:40.099 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:40.099 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:40.109 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117443 files and directories currently installed.) 2026-03-10T10:22:40.111 INFO:teuthology.orchestra.run.vm00.stdout:Removing ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:40.211 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T10:22:40.211 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 3673 kB disk space will be freed. 2026-03-10T10:22:40.243 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:40.243 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:40.243 INFO:teuthology.orchestra.run.vm03.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T10:22:40.244 INFO:teuthology.orchestra.run.vm03.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T10:22:40.244 INFO:teuthology.orchestra.run.vm03.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:22:40.244 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T10:22:40.244 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:22:40.244 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:22:40.244 INFO:teuthology.orchestra.run.vm03.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:22:40.244 INFO:teuthology.orchestra.run.vm03.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T10:22:40.244 INFO:teuthology.orchestra.run.vm03.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T10:22:40.244 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:22:40.244 INFO:teuthology.orchestra.run.vm03.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:22:40.244 INFO:teuthology.orchestra.run.vm03.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:22:40.244 INFO:teuthology.orchestra.run.vm03.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T10:22:40.244 INFO:teuthology.orchestra.run.vm03.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:40.244 INFO:teuthology.orchestra.run.vm03.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T10:22:40.244 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:40.251 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117443 files and directories currently installed.) 2026-03-10T10:22:40.253 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:40.254 INFO:teuthology.orchestra.run.vm03.stdout:The following packages will be REMOVED: 2026-03-10T10:22:40.255 INFO:teuthology.orchestra.run.vm03.stdout: ceph-fuse* 2026-03-10T10:22:40.424 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T10:22:40.424 INFO:teuthology.orchestra.run.vm03.stdout:After this operation, 3673 kB disk space will be freed. 2026-03-10T10:22:40.462 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117443 files and directories currently installed.) 2026-03-10T10:22:40.465 INFO:teuthology.orchestra.run.vm03.stdout:Removing ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:40.491 INFO:teuthology.orchestra.run.vm00.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:40.575 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117434 files and directories currently installed.) 2026-03-10T10:22:40.577 INFO:teuthology.orchestra.run.vm00.stdout:Purging configuration files for ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:40.646 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:40.738 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117434 files and directories currently installed.) 2026-03-10T10:22:40.740 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:40.886 INFO:teuthology.orchestra.run.vm03.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:40.984 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117434 files and directories currently installed.) 2026-03-10T10:22:40.987 INFO:teuthology.orchestra.run.vm03.stdout:Purging configuration files for ceph-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:42.058 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:42.091 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:42.275 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:42.276 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:42.316 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:42.349 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout:Package 'ceph-test' is not installed, so not removed 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T10:22:42.395 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:42.410 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:42.410 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:42.442 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:42.479 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:42.515 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:42.548 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:42.548 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:42.633 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:42.634 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:42.693 INFO:teuthology.orchestra.run.vm09.stdout:Package 'ceph-test' is not installed, so not removed 2026-03-10T10:22:42.693 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:42.693 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:42.693 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T10:22:42.694 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T10:22:42.694 INFO:teuthology.orchestra.run.vm09.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:22:42.694 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T10:22:42.694 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:22:42.694 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:22:42.694 INFO:teuthology.orchestra.run.vm09.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:22:42.694 INFO:teuthology.orchestra.run.vm09.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T10:22:42.694 INFO:teuthology.orchestra.run.vm09.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T10:22:42.694 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:22:42.694 INFO:teuthology.orchestra.run.vm09.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:22:42.694 INFO:teuthology.orchestra.run.vm09.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:22:42.694 INFO:teuthology.orchestra.run.vm09.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T10:22:42.694 INFO:teuthology.orchestra.run.vm09.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:42.694 INFO:teuthology.orchestra.run.vm09.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T10:22:42.694 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:42.719 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:42.720 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:42.720 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:42.720 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:42.752 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:42.851 INFO:teuthology.orchestra.run.vm00.stdout:Package 'ceph-volume' is not installed, so not removed 2026-03-10T10:22:42.851 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:42.851 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:42.852 INFO:teuthology.orchestra.run.vm00.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T10:22:42.852 INFO:teuthology.orchestra.run.vm00.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T10:22:42.852 INFO:teuthology.orchestra.run.vm00.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:22:42.852 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T10:22:42.852 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:22:42.852 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:22:42.852 INFO:teuthology.orchestra.run.vm00.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:22:42.852 INFO:teuthology.orchestra.run.vm00.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T10:22:42.853 INFO:teuthology.orchestra.run.vm00.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T10:22:42.853 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:22:42.853 INFO:teuthology.orchestra.run.vm00.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:22:42.853 INFO:teuthology.orchestra.run.vm00.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:22:42.853 INFO:teuthology.orchestra.run.vm00.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T10:22:42.853 INFO:teuthology.orchestra.run.vm00.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:42.853 INFO:teuthology.orchestra.run.vm00.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T10:22:42.853 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:42.875 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:42.875 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:42.911 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:42.935 INFO:teuthology.orchestra.run.vm03.stdout:Package 'ceph-test' is not installed, so not removed 2026-03-10T10:22:42.935 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:42.935 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:42.935 INFO:teuthology.orchestra.run.vm03.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T10:22:42.936 INFO:teuthology.orchestra.run.vm03.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T10:22:42.936 INFO:teuthology.orchestra.run.vm03.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:22:42.936 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T10:22:42.936 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:22:42.936 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:22:42.936 INFO:teuthology.orchestra.run.vm03.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:22:42.936 INFO:teuthology.orchestra.run.vm03.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T10:22:42.936 INFO:teuthology.orchestra.run.vm03.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T10:22:42.936 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:22:42.936 INFO:teuthology.orchestra.run.vm03.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:22:42.936 INFO:teuthology.orchestra.run.vm03.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:22:42.936 INFO:teuthology.orchestra.run.vm03.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T10:22:42.936 INFO:teuthology.orchestra.run.vm03.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:42.936 INFO:teuthology.orchestra.run.vm03.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T10:22:42.936 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:42.960 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:42.960 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:42.970 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:42.970 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:42.994 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:43.108 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:43.108 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:43.181 INFO:teuthology.orchestra.run.vm09.stdout:Package 'ceph-volume' is not installed, so not removed 2026-03-10T10:22:43.181 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:43.181 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:43.181 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T10:22:43.182 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T10:22:43.182 INFO:teuthology.orchestra.run.vm09.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:22:43.182 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T10:22:43.182 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:22:43.182 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:22:43.182 INFO:teuthology.orchestra.run.vm09.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:22:43.182 INFO:teuthology.orchestra.run.vm09.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T10:22:43.182 INFO:teuthology.orchestra.run.vm09.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T10:22:43.182 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:22:43.182 INFO:teuthology.orchestra.run.vm09.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:22:43.182 INFO:teuthology.orchestra.run.vm09.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:22:43.182 INFO:teuthology.orchestra.run.vm09.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T10:22:43.182 INFO:teuthology.orchestra.run.vm09.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:43.182 INFO:teuthology.orchestra.run.vm09.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T10:22:43.182 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:43.206 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:43.206 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:43.216 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:43.217 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:43.238 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:43.317 INFO:teuthology.orchestra.run.vm00.stdout:Package 'radosgw' is not installed, so not removed 2026-03-10T10:22:43.317 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:43.317 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:43.317 INFO:teuthology.orchestra.run.vm00.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T10:22:43.318 INFO:teuthology.orchestra.run.vm00.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T10:22:43.318 INFO:teuthology.orchestra.run.vm00.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:22:43.318 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T10:22:43.318 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:22:43.318 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:22:43.318 INFO:teuthology.orchestra.run.vm00.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:22:43.318 INFO:teuthology.orchestra.run.vm00.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T10:22:43.318 INFO:teuthology.orchestra.run.vm00.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T10:22:43.318 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:22:43.318 INFO:teuthology.orchestra.run.vm00.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:22:43.318 INFO:teuthology.orchestra.run.vm00.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:22:43.319 INFO:teuthology.orchestra.run.vm00.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T10:22:43.319 INFO:teuthology.orchestra.run.vm00.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:43.319 INFO:teuthology.orchestra.run.vm00.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T10:22:43.319 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:43.344 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:43.344 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:43.378 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:43.419 INFO:teuthology.orchestra.run.vm03.stdout:Package 'ceph-volume' is not installed, so not removed 2026-03-10T10:22:43.419 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:43.420 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:43.420 INFO:teuthology.orchestra.run.vm03.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T10:22:43.420 INFO:teuthology.orchestra.run.vm03.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T10:22:43.420 INFO:teuthology.orchestra.run.vm03.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:22:43.420 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T10:22:43.420 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:22:43.421 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:22:43.421 INFO:teuthology.orchestra.run.vm03.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:22:43.421 INFO:teuthology.orchestra.run.vm03.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T10:22:43.421 INFO:teuthology.orchestra.run.vm03.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T10:22:43.421 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:22:43.421 INFO:teuthology.orchestra.run.vm03.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:22:43.421 INFO:teuthology.orchestra.run.vm03.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:22:43.421 INFO:teuthology.orchestra.run.vm03.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T10:22:43.421 INFO:teuthology.orchestra.run.vm03.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:43.421 INFO:teuthology.orchestra.run.vm03.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T10:22:43.421 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:43.438 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:43.438 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:43.456 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:43.456 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:43.470 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:43.572 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:43.572 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:43.622 INFO:teuthology.orchestra.run.vm09.stdout:Package 'radosgw' is not installed, so not removed 2026-03-10T10:22:43.622 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:43.622 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:43.622 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T10:22:43.622 INFO:teuthology.orchestra.run.vm09.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T10:22:43.622 INFO:teuthology.orchestra.run.vm09.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:22:43.622 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T10:22:43.622 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:22:43.622 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:22:43.623 INFO:teuthology.orchestra.run.vm09.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:22:43.623 INFO:teuthology.orchestra.run.vm09.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T10:22:43.623 INFO:teuthology.orchestra.run.vm09.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T10:22:43.623 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:22:43.623 INFO:teuthology.orchestra.run.vm09.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:22:43.623 INFO:teuthology.orchestra.run.vm09.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:22:43.623 INFO:teuthology.orchestra.run.vm09.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T10:22:43.623 INFO:teuthology.orchestra.run.vm09.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:43.623 INFO:teuthology.orchestra.run.vm09.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T10:22:43.623 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:43.640 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:43.640 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:43.672 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:43.689 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:43.690 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:43.692 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:43.692 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:43.692 INFO:teuthology.orchestra.run.vm00.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:43.692 INFO:teuthology.orchestra.run.vm00.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:43.693 INFO:teuthology.orchestra.run.vm00.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:43.693 INFO:teuthology.orchestra.run.vm00.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:43.693 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:43.693 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:43.693 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:43.693 INFO:teuthology.orchestra.run.vm00.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:43.693 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:43.693 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:43.693 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:43.693 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:43.693 INFO:teuthology.orchestra.run.vm00.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:43.693 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:43.693 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:43.693 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:43.693 INFO:teuthology.orchestra.run.vm00.stdout: xmlstarlet zip 2026-03-10T10:22:43.693 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:43.710 INFO:teuthology.orchestra.run.vm00.stdout:The following packages will be REMOVED: 2026-03-10T10:22:43.710 INFO:teuthology.orchestra.run.vm00.stdout: python3-cephfs* python3-rados* python3-rgw* 2026-03-10T10:22:43.882 INFO:teuthology.orchestra.run.vm03.stdout:Package 'radosgw' is not installed, so not removed 2026-03-10T10:22:43.882 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:43.883 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:43.883 INFO:teuthology.orchestra.run.vm03.stdout: libboost-thread1.74.0 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-10T10:22:43.883 INFO:teuthology.orchestra.run.vm03.stdout: libsgutils2-2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-10T10:22:43.883 INFO:teuthology.orchestra.run.vm03.stdout: python-pastedeploy-tpl python3-asyncssh python3-cachetools 2026-03-10T10:22:43.883 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-10T10:22:43.884 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-10T10:22:43.884 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-logutils 2026-03-10T10:22:43.884 INFO:teuthology.orchestra.run.vm03.stdout: python3-mako python3-natsort python3-paste python3-pastedeploy 2026-03-10T10:22:43.884 INFO:teuthology.orchestra.run.vm03.stdout: python3-pastescript python3-pecan python3-portend python3-prettytable 2026-03-10T10:22:43.884 INFO:teuthology.orchestra.run.vm03.stdout: python3-psutil python3-pyinotify python3-repoze.lru 2026-03-10T10:22:43.884 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplegeneric 2026-03-10T10:22:43.884 INFO:teuthology.orchestra.run.vm03.stdout: python3-simplejson python3-singledispatch python3-sklearn 2026-03-10T10:22:43.884 INFO:teuthology.orchestra.run.vm03.stdout: python3-sklearn-lib python3-tempita python3-tempora python3-threadpoolctl 2026-03-10T10:22:43.884 INFO:teuthology.orchestra.run.vm03.stdout: python3-waitress python3-wcwidth python3-webob python3-websocket 2026-03-10T10:22:43.884 INFO:teuthology.orchestra.run.vm03.stdout: python3-webtest python3-werkzeug python3-zc.lockfile sg3-utils 2026-03-10T10:22:43.884 INFO:teuthology.orchestra.run.vm03.stdout: sg3-utils-udev smartmontools socat xmlstarlet 2026-03-10T10:22:43.884 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:43.886 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:43.887 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:43.905 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 3 to remove and 10 not upgraded. 2026-03-10T10:22:43.906 INFO:teuthology.orchestra.run.vm00.stdout:After this operation, 2062 kB disk space will be freed. 2026-03-10T10:22:43.908 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:43.908 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:43.943 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:43.948 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117434 files and directories currently installed.) 2026-03-10T10:22:43.950 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:43.962 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:43.973 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:44.037 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:44.037 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:44.038 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:44.038 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:44.038 INFO:teuthology.orchestra.run.vm09.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:44.038 INFO:teuthology.orchestra.run.vm09.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:44.038 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:44.038 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:44.039 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:44.039 INFO:teuthology.orchestra.run.vm09.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:44.039 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:44.039 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:44.039 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:44.039 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:44.039 INFO:teuthology.orchestra.run.vm09.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:44.039 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:44.039 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:44.039 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:44.039 INFO:teuthology.orchestra.run.vm09.stdout: xmlstarlet zip 2026-03-10T10:22:44.039 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:44.057 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-10T10:22:44.058 INFO:teuthology.orchestra.run.vm09.stdout: python3-cephfs* python3-rados* python3-rgw* 2026-03-10T10:22:44.140 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:44.140 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:44.235 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 3 to remove and 10 not upgraded. 2026-03-10T10:22:44.235 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 2062 kB disk space will be freed. 2026-03-10T10:22:44.269 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117434 files and directories currently installed.) 2026-03-10T10:22:44.271 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:44.273 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:44.273 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:44.273 INFO:teuthology.orchestra.run.vm03.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:44.273 INFO:teuthology.orchestra.run.vm03.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:44.274 INFO:teuthology.orchestra.run.vm03.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:44.274 INFO:teuthology.orchestra.run.vm03.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:44.274 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:44.274 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:44.274 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:44.274 INFO:teuthology.orchestra.run.vm03.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:44.274 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:44.274 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:44.274 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:44.274 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:44.274 INFO:teuthology.orchestra.run.vm03.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:44.274 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:44.274 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:44.274 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:44.274 INFO:teuthology.orchestra.run.vm03.stdout: xmlstarlet zip 2026-03-10T10:22:44.274 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:44.282 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:44.284 INFO:teuthology.orchestra.run.vm03.stdout:The following packages will be REMOVED: 2026-03-10T10:22:44.284 INFO:teuthology.orchestra.run.vm03.stdout: python3-cephfs* python3-rados* python3-rgw* 2026-03-10T10:22:44.294 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:44.452 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 3 to remove and 10 not upgraded. 2026-03-10T10:22:44.452 INFO:teuthology.orchestra.run.vm03.stdout:After this operation, 2062 kB disk space will be freed. 2026-03-10T10:22:44.484 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117434 files and directories currently installed.) 2026-03-10T10:22:44.485 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-cephfs (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:44.495 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-rgw (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:44.535 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-rados (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:45.042 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:45.075 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:45.288 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:45.289 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:45.381 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout:Package 'python3-rgw' is not installed, so not removed 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout: xmlstarlet zip 2026-03-10T10:22:45.386 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:45.400 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:45.400 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:45.416 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:45.434 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:45.584 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:45.584 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:45.606 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:45.633 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:45.634 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:45.641 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:45.803 INFO:teuthology.orchestra.run.vm00.stdout:Package 'python3-cephfs' is not installed, so not removed 2026-03-10T10:22:45.803 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:45.803 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:45.804 INFO:teuthology.orchestra.run.vm00.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:45.804 INFO:teuthology.orchestra.run.vm00.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:45.804 INFO:teuthology.orchestra.run.vm00.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:45.804 INFO:teuthology.orchestra.run.vm00.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:45.804 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:45.805 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:45.805 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:45.805 INFO:teuthology.orchestra.run.vm00.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:45.805 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:45.805 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:45.805 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:45.805 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:45.805 INFO:teuthology.orchestra.run.vm00.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:45.805 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:45.805 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:45.805 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:45.805 INFO:teuthology.orchestra.run.vm00.stdout: xmlstarlet zip 2026-03-10T10:22:45.805 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:45.830 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:45.831 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:45.855 INFO:teuthology.orchestra.run.vm09.stdout:Package 'python3-rgw' is not installed, so not removed 2026-03-10T10:22:45.855 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:45.855 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:45.855 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:45.855 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:45.855 INFO:teuthology.orchestra.run.vm09.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:45.855 INFO:teuthology.orchestra.run.vm09.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:45.856 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:45.856 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:45.856 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:45.856 INFO:teuthology.orchestra.run.vm09.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:45.856 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:45.856 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:45.856 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:45.856 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:45.856 INFO:teuthology.orchestra.run.vm09.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:45.856 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:45.856 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:45.856 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:45.856 INFO:teuthology.orchestra.run.vm09.stdout: xmlstarlet zip 2026-03-10T10:22:45.856 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:45.862 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:45.863 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:45.864 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:45.887 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:45.887 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:45.920 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:46.009 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:46.009 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:46.037 INFO:teuthology.orchestra.run.vm03.stdout:Package 'python3-rgw' is not installed, so not removed 2026-03-10T10:22:46.037 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:46.037 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:46.037 INFO:teuthology.orchestra.run.vm03.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:46.037 INFO:teuthology.orchestra.run.vm03.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:46.038 INFO:teuthology.orchestra.run.vm03.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:46.038 INFO:teuthology.orchestra.run.vm03.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:46.038 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:46.038 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:46.038 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:46.038 INFO:teuthology.orchestra.run.vm03.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:46.038 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:46.038 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:46.038 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:46.038 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:46.038 INFO:teuthology.orchestra.run.vm03.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:46.038 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:46.038 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:46.038 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:46.038 INFO:teuthology.orchestra.run.vm03.stdout: xmlstarlet zip 2026-03-10T10:22:46.038 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:46.066 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:46.066 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:46.099 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:46.131 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:46.134 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:46.211 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:46.211 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:46.211 INFO:teuthology.orchestra.run.vm00.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:46.211 INFO:teuthology.orchestra.run.vm00.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:46.212 INFO:teuthology.orchestra.run.vm00.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:46.212 INFO:teuthology.orchestra.run.vm00.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:46.212 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:46.212 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:46.212 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:46.212 INFO:teuthology.orchestra.run.vm00.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:46.212 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:46.212 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:46.212 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:46.212 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:46.212 INFO:teuthology.orchestra.run.vm00.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:46.212 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:46.212 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:46.212 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:46.212 INFO:teuthology.orchestra.run.vm00.stdout: xmlstarlet zip 2026-03-10T10:22:46.212 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:46.228 INFO:teuthology.orchestra.run.vm00.stdout:The following packages will be REMOVED: 2026-03-10T10:22:46.228 INFO:teuthology.orchestra.run.vm00.stdout: python3-rbd* 2026-03-10T10:22:46.310 INFO:teuthology.orchestra.run.vm09.stdout:Package 'python3-cephfs' is not installed, so not removed 2026-03-10T10:22:46.310 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:46.310 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:46.311 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:46.311 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:46.311 INFO:teuthology.orchestra.run.vm09.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:46.311 INFO:teuthology.orchestra.run.vm09.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:46.311 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:46.311 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:46.311 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:46.311 INFO:teuthology.orchestra.run.vm09.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:46.311 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:46.311 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:46.311 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:46.311 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:46.311 INFO:teuthology.orchestra.run.vm09.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:46.311 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:46.312 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:46.312 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:46.312 INFO:teuthology.orchestra.run.vm09.stdout: xmlstarlet zip 2026-03-10T10:22:46.312 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:46.318 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:46.319 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:46.337 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:46.337 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:46.368 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:46.412 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T10:22:46.412 INFO:teuthology.orchestra.run.vm00.stdout:After this operation, 1186 kB disk space will be freed. 2026-03-10T10:22:46.452 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117410 files and directories currently installed.) 2026-03-10T10:22:46.454 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:46.493 INFO:teuthology.orchestra.run.vm03.stdout:Package 'python3-cephfs' is not installed, so not removed 2026-03-10T10:22:46.493 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:46.493 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:46.493 INFO:teuthology.orchestra.run.vm03.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:46.493 INFO:teuthology.orchestra.run.vm03.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:46.494 INFO:teuthology.orchestra.run.vm03.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:46.495 INFO:teuthology.orchestra.run.vm03.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:46.495 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:46.495 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:46.495 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:46.495 INFO:teuthology.orchestra.run.vm03.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:46.495 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:46.495 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:46.495 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:46.495 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:46.495 INFO:teuthology.orchestra.run.vm03.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:46.495 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:46.495 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:46.495 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:46.495 INFO:teuthology.orchestra.run.vm03.stdout: xmlstarlet zip 2026-03-10T10:22:46.495 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:46.517 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:46.518 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:46.553 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:46.576 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:46.577 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:46.742 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:46.742 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:46.742 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:46.742 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:46.743 INFO:teuthology.orchestra.run.vm09.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:46.743 INFO:teuthology.orchestra.run.vm09.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:46.743 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:46.743 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:46.743 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:46.743 INFO:teuthology.orchestra.run.vm09.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:46.743 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:46.743 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:46.743 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:46.743 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:46.743 INFO:teuthology.orchestra.run.vm09.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:46.743 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:46.743 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:46.743 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:46.743 INFO:teuthology.orchestra.run.vm09.stdout: xmlstarlet zip 2026-03-10T10:22:46.743 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:46.755 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-10T10:22:46.755 INFO:teuthology.orchestra.run.vm09.stdout: python3-rbd* 2026-03-10T10:22:46.773 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:46.774 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:46.926 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T10:22:46.926 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 1186 kB disk space will be freed. 2026-03-10T10:22:46.961 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117410 files and directories currently installed.) 2026-03-10T10:22:46.964 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:46.973 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:46.973 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:46.973 INFO:teuthology.orchestra.run.vm03.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:46.973 INFO:teuthology.orchestra.run.vm03.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:46.974 INFO:teuthology.orchestra.run.vm03.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:46.974 INFO:teuthology.orchestra.run.vm03.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:46.974 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:46.974 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:46.974 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:46.974 INFO:teuthology.orchestra.run.vm03.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:46.974 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:46.974 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:46.974 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:46.974 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:46.974 INFO:teuthology.orchestra.run.vm03.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:46.975 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:46.975 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:46.975 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:46.975 INFO:teuthology.orchestra.run.vm03.stdout: xmlstarlet zip 2026-03-10T10:22:46.975 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:46.991 INFO:teuthology.orchestra.run.vm03.stdout:The following packages will be REMOVED: 2026-03-10T10:22:46.991 INFO:teuthology.orchestra.run.vm03.stdout: python3-rbd* 2026-03-10T10:22:47.186 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 1 to remove and 10 not upgraded. 2026-03-10T10:22:47.187 INFO:teuthology.orchestra.run.vm03.stdout:After this operation, 1186 kB disk space will be freed. 2026-03-10T10:22:47.228 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117410 files and directories currently installed.) 2026-03-10T10:22:47.230 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-rbd (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:47.501 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:47.534 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:47.684 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:47.684 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:47.867 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:47.867 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:47.867 INFO:teuthology.orchestra.run.vm00.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:47.867 INFO:teuthology.orchestra.run.vm00.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:47.868 INFO:teuthology.orchestra.run.vm00.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:47.868 INFO:teuthology.orchestra.run.vm00.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:47.868 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:47.868 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:47.868 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:47.868 INFO:teuthology.orchestra.run.vm00.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:47.868 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:47.868 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:47.868 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:47.868 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:47.868 INFO:teuthology.orchestra.run.vm00.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:47.868 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:47.868 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:47.868 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:47.868 INFO:teuthology.orchestra.run.vm00.stdout: xmlstarlet zip 2026-03-10T10:22:47.868 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:47.883 INFO:teuthology.orchestra.run.vm00.stdout:The following packages will be REMOVED: 2026-03-10T10:22:47.884 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs-dev* libcephfs2* 2026-03-10T10:22:48.062 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 2 to remove and 10 not upgraded. 2026-03-10T10:22:48.062 INFO:teuthology.orchestra.run.vm00.stdout:After this operation, 3202 kB disk space will be freed. 2026-03-10T10:22:48.102 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:48.103 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117402 files and directories currently installed.) 2026-03-10T10:22:48.106 INFO:teuthology.orchestra.run.vm00.stdout:Removing libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:48.117 INFO:teuthology.orchestra.run.vm00.stdout:Removing libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:48.134 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:48.142 INFO:teuthology.orchestra.run.vm00.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T10:22:48.349 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:48.349 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:48.397 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:48.430 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:48.518 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:48.518 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:48.518 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:48.518 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:48.519 INFO:teuthology.orchestra.run.vm09.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:48.519 INFO:teuthology.orchestra.run.vm09.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:48.519 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:48.519 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:48.519 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:48.519 INFO:teuthology.orchestra.run.vm09.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:48.519 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:48.519 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:48.519 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:48.519 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:48.519 INFO:teuthology.orchestra.run.vm09.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:48.519 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:48.519 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:48.520 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:48.520 INFO:teuthology.orchestra.run.vm09.stdout: xmlstarlet zip 2026-03-10T10:22:48.520 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:48.532 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-10T10:22:48.533 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-dev* libcephfs2* 2026-03-10T10:22:48.626 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:48.627 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:48.713 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 2 to remove and 10 not upgraded. 2026-03-10T10:22:48.713 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 3202 kB disk space will be freed. 2026-03-10T10:22:48.752 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117402 files and directories currently installed.) 2026-03-10T10:22:48.755 INFO:teuthology.orchestra.run.vm09.stdout:Removing libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:48.766 INFO:teuthology.orchestra.run.vm09.stdout:Removing libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:48.784 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:48.784 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:48.785 INFO:teuthology.orchestra.run.vm03.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:48.785 INFO:teuthology.orchestra.run.vm03.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:48.785 INFO:teuthology.orchestra.run.vm03.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:48.785 INFO:teuthology.orchestra.run.vm03.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:48.785 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:48.785 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:48.786 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:48.786 INFO:teuthology.orchestra.run.vm03.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:48.786 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:48.786 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:48.786 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:48.786 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:48.786 INFO:teuthology.orchestra.run.vm03.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:48.786 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:48.786 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:48.786 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:48.786 INFO:teuthology.orchestra.run.vm03.stdout: xmlstarlet zip 2026-03-10T10:22:48.786 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:48.792 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T10:22:48.799 INFO:teuthology.orchestra.run.vm03.stdout:The following packages will be REMOVED: 2026-03-10T10:22:48.800 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs-dev* libcephfs2* 2026-03-10T10:22:48.979 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 2 to remove and 10 not upgraded. 2026-03-10T10:22:48.979 INFO:teuthology.orchestra.run.vm03.stdout:After this operation, 3202 kB disk space will be freed. 2026-03-10T10:22:49.018 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117402 files and directories currently installed.) 2026-03-10T10:22:49.021 INFO:teuthology.orchestra.run.vm03.stdout:Removing libcephfs-dev (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:49.033 INFO:teuthology.orchestra.run.vm03.stdout:Removing libcephfs2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:49.059 INFO:teuthology.orchestra.run.vm03.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T10:22:49.174 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:49.207 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:49.396 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:49.396 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout:Package 'libcephfs-dev' is not installed, so not removed 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout: xmlstarlet zip 2026-03-10T10:22:49.542 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:49.561 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:49.561 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:49.595 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:49.775 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:49.775 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:49.888 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:49.923 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T10:22:49.952 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:49.961 INFO:teuthology.orchestra.run.vm00.stdout:The following packages will be REMOVED: 2026-03-10T10:22:49.961 INFO:teuthology.orchestra.run.vm00.stdout: librados2* libradosstriper1* librbd1* librgw2* libsqlite3-mod-ceph* 2026-03-10T10:22:49.961 INFO:teuthology.orchestra.run.vm00.stdout: qemu-block-extra* rbd-fuse* 2026-03-10T10:22:50.112 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:50.112 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:50.121 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 7 to remove and 10 not upgraded. 2026-03-10T10:22:50.121 INFO:teuthology.orchestra.run.vm00.stdout:After this operation, 51.6 MB disk space will be freed. 2026-03-10T10:22:50.128 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:50.153 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117387 files and directories currently installed.) 2026-03-10T10:22:50.154 INFO:teuthology.orchestra.run.vm00.stdout:Removing rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:50.161 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:50.166 INFO:teuthology.orchestra.run.vm00.stdout:Removing libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:50.177 INFO:teuthology.orchestra.run.vm00.stdout:Removing libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:50.187 INFO:teuthology.orchestra.run.vm00.stdout:Removing qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-10T10:22:50.241 INFO:teuthology.orchestra.run.vm09.stdout:Package 'libcephfs-dev' is not installed, so not removed 2026-03-10T10:22:50.241 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:50.241 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:50.241 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:50.241 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:50.242 INFO:teuthology.orchestra.run.vm09.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:50.242 INFO:teuthology.orchestra.run.vm09.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:50.242 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:50.242 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:50.242 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:50.242 INFO:teuthology.orchestra.run.vm09.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:50.242 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:50.242 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:50.242 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:50.242 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:50.242 INFO:teuthology.orchestra.run.vm09.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:50.242 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:50.242 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:50.242 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:50.242 INFO:teuthology.orchestra.run.vm09.stdout: xmlstarlet zip 2026-03-10T10:22:50.242 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:50.259 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:50.260 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:50.291 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:50.352 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:50.353 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:50.467 INFO:teuthology.orchestra.run.vm03.stdout:Package 'libcephfs-dev' is not installed, so not removed 2026-03-10T10:22:50.467 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:50.467 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:50.467 INFO:teuthology.orchestra.run.vm03.stdout: libboost-thread1.74.0 libjq1 liblua5.3-dev liboath0 libonig5 libpmemobj1 2026-03-10T10:22:50.468 INFO:teuthology.orchestra.run.vm03.stdout: libradosstriper1 librdkafka1 libreadline-dev librgw2 libsgutils2-2 2026-03-10T10:22:50.468 INFO:teuthology.orchestra.run.vm03.stdout: libsqlite3-mod-ceph lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli 2026-03-10T10:22:50.468 INFO:teuthology.orchestra.run.vm03.stdout: pkg-config python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:50.468 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:50.468 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:50.468 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:50.468 INFO:teuthology.orchestra.run.vm03.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:50.468 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:50.468 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:50.468 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:50.468 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:50.468 INFO:teuthology.orchestra.run.vm03.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:50.468 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:50.468 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:50.468 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc.lockfile sg3-utils sg3-utils-udev smartmontools socat unzip 2026-03-10T10:22:50.468 INFO:teuthology.orchestra.run.vm03.stdout: xmlstarlet zip 2026-03-10T10:22:50.468 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:50.477 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:50.477 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:50.484 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:50.484 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:50.515 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:50.603 INFO:teuthology.orchestra.run.vm00.stdout:Removing librbd1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:50.609 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:50.609 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:50.609 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T10:22:50.610 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:50.616 INFO:teuthology.orchestra.run.vm00.stdout:Removing librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:50.622 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-10T10:22:50.623 INFO:teuthology.orchestra.run.vm09.stdout: librados2* libradosstriper1* librbd1* librgw2* libsqlite3-mod-ceph* 2026-03-10T10:22:50.623 INFO:teuthology.orchestra.run.vm09.stdout: qemu-block-extra* rbd-fuse* 2026-03-10T10:22:50.629 INFO:teuthology.orchestra.run.vm00.stdout:Removing librados2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:50.654 INFO:teuthology.orchestra.run.vm00.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:50.690 INFO:teuthology.orchestra.run.vm00.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T10:22:50.700 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:50.700 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:50.761 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117336 files and directories currently installed.) 2026-03-10T10:22:50.763 INFO:teuthology.orchestra.run.vm00.stdout:Purging configuration files for qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-10T10:22:50.788 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 7 to remove and 10 not upgraded. 2026-03-10T10:22:50.788 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 51.6 MB disk space will be freed. 2026-03-10T10:22:50.820 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:50.820 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:50.820 INFO:teuthology.orchestra.run.vm03.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T10:22:50.820 INFO:teuthology.orchestra.run.vm03.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T10:22:50.820 INFO:teuthology.orchestra.run.vm03.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T10:22:50.821 INFO:teuthology.orchestra.run.vm03.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T10:22:50.821 INFO:teuthology.orchestra.run.vm03.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T10:22:50.821 INFO:teuthology.orchestra.run.vm03.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:50.821 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:50.821 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:50.821 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:50.821 INFO:teuthology.orchestra.run.vm03.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:50.821 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:50.821 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:50.821 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:50.821 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:50.821 INFO:teuthology.orchestra.run.vm03.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:50.821 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:50.821 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:50.821 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T10:22:50.821 INFO:teuthology.orchestra.run.vm03.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T10:22:50.821 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:50.824 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117387 files and directories currently installed.) 2026-03-10T10:22:50.827 INFO:teuthology.orchestra.run.vm09.stdout:Removing rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:50.831 INFO:teuthology.orchestra.run.vm03.stdout:The following packages will be REMOVED: 2026-03-10T10:22:50.831 INFO:teuthology.orchestra.run.vm03.stdout: librados2* libradosstriper1* librbd1* librgw2* libsqlite3-mod-ceph* 2026-03-10T10:22:50.831 INFO:teuthology.orchestra.run.vm03.stdout: qemu-block-extra* rbd-fuse* 2026-03-10T10:22:50.839 INFO:teuthology.orchestra.run.vm09.stdout:Removing libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:50.851 INFO:teuthology.orchestra.run.vm09.stdout:Removing libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:50.863 INFO:teuthology.orchestra.run.vm09.stdout:Removing qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-10T10:22:50.993 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 7 to remove and 10 not upgraded. 2026-03-10T10:22:50.993 INFO:teuthology.orchestra.run.vm03.stdout:After this operation, 51.6 MB disk space will be freed. 2026-03-10T10:22:51.028 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117387 files and directories currently installed.) 2026-03-10T10:22:51.030 INFO:teuthology.orchestra.run.vm03.stdout:Removing rbd-fuse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:51.043 INFO:teuthology.orchestra.run.vm03.stdout:Removing libsqlite3-mod-ceph (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:51.056 INFO:teuthology.orchestra.run.vm03.stdout:Removing libradosstriper1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:51.068 INFO:teuthology.orchestra.run.vm03.stdout:Removing qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-10T10:22:51.277 INFO:teuthology.orchestra.run.vm09.stdout:Removing librbd1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:51.289 INFO:teuthology.orchestra.run.vm09.stdout:Removing librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:51.301 INFO:teuthology.orchestra.run.vm09.stdout:Removing librados2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:51.328 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:51.361 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T10:22:51.423 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117336 files and directories currently installed.) 2026-03-10T10:22:51.424 INFO:teuthology.orchestra.run.vm09.stdout:Purging configuration files for qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-10T10:22:51.495 INFO:teuthology.orchestra.run.vm03.stdout:Removing librbd1 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:51.508 INFO:teuthology.orchestra.run.vm03.stdout:Removing librgw2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:51.522 INFO:teuthology.orchestra.run.vm03.stdout:Removing librados2 (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:51.549 INFO:teuthology.orchestra.run.vm03.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:51.588 INFO:teuthology.orchestra.run.vm03.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T10:22:51.656 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117336 files and directories currently installed.) 2026-03-10T10:22:51.658 INFO:teuthology.orchestra.run.vm03.stdout:Purging configuration files for qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-10T10:22:52.163 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:52.197 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:52.394 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:52.394 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout:Package 'librbd1' is not installed, so not removed 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T10:22:52.587 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:52.610 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:52.610 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:52.640 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:52.838 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:52.839 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:52.955 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:52.989 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:52.989 INFO:teuthology.orchestra.run.vm00.stdout:Package 'rbd-fuse' is not installed, so not removed 2026-03-10T10:22:52.989 INFO:teuthology.orchestra.run.vm00.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:52.989 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:52.989 INFO:teuthology.orchestra.run.vm00.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T10:22:52.989 INFO:teuthology.orchestra.run.vm00.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T10:22:52.989 INFO:teuthology.orchestra.run.vm00.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T10:22:52.990 INFO:teuthology.orchestra.run.vm00.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T10:22:52.990 INFO:teuthology.orchestra.run.vm00.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T10:22:52.990 INFO:teuthology.orchestra.run.vm00.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:52.990 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:52.990 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:52.990 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:52.990 INFO:teuthology.orchestra.run.vm00.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:52.990 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:52.990 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:52.990 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:52.991 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:52.991 INFO:teuthology.orchestra.run.vm00.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:52.991 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:52.991 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:52.991 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T10:22:52.991 INFO:teuthology.orchestra.run.vm00.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T10:22:52.991 INFO:teuthology.orchestra.run.vm00.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:53.020 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:53.020 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:53.022 DEBUG:teuthology.orchestra.run.vm00:> dpkg -l | grep '^.\(U\|H\)R' | awk '{print $2}' | sudo xargs --no-run-if-empty dpkg -P --force-remove-reinstreq 2026-03-10T10:22:53.079 DEBUG:teuthology.orchestra.run.vm00:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" autoremove 2026-03-10T10:22:53.151 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:22:53.162 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:53.171 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:53.172 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:53.198 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:53.313 INFO:teuthology.orchestra.run.vm09.stdout:Package 'librbd1' is not installed, so not removed 2026-03-10T10:22:53.313 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:53.313 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:53.313 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T10:22:53.313 INFO:teuthology.orchestra.run.vm09.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T10:22:53.313 INFO:teuthology.orchestra.run.vm09.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T10:22:53.313 INFO:teuthology.orchestra.run.vm09.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T10:22:53.313 INFO:teuthology.orchestra.run.vm09.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T10:22:53.313 INFO:teuthology.orchestra.run.vm09.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:53.313 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:53.313 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:53.313 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:53.313 INFO:teuthology.orchestra.run.vm09.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:53.313 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:53.314 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:53.314 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:53.314 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:53.314 INFO:teuthology.orchestra.run.vm09.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:53.314 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:53.314 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:53.314 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T10:22:53.314 INFO:teuthology.orchestra.run.vm09.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T10:22:53.314 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:53.334 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:53.335 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:53.356 INFO:teuthology.orchestra.run.vm00.stdout:Building dependency tree... 2026-03-10T10:22:53.357 INFO:teuthology.orchestra.run.vm00.stdout:Reading state information... 2026-03-10T10:22:53.367 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:53.395 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:53.395 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:53.537 INFO:teuthology.orchestra.run.vm00.stdout:The following packages will be REMOVED: 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T10:22:53.538 INFO:teuthology.orchestra.run.vm00.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T10:22:53.565 INFO:teuthology.orchestra.run.vm03.stdout:Package 'librbd1' is not installed, so not removed 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:53.566 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:53.567 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T10:22:53.567 INFO:teuthology.orchestra.run.vm03.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T10:22:53.567 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:53.571 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:53.572 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:53.586 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:53.586 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:53.618 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:53.720 INFO:teuthology.orchestra.run.vm00.stdout:0 upgraded, 0 newly installed, 87 to remove and 10 not upgraded. 2026-03-10T10:22:53.720 INFO:teuthology.orchestra.run.vm00.stdout:After this operation, 107 MB disk space will be freed. 2026-03-10T10:22:53.721 INFO:teuthology.orchestra.run.vm09.stdout:Package 'rbd-fuse' is not installed, so not removed 2026-03-10T10:22:53.721 INFO:teuthology.orchestra.run.vm09.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:53.722 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:53.722 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T10:22:53.722 INFO:teuthology.orchestra.run.vm09.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T10:22:53.722 INFO:teuthology.orchestra.run.vm09.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T10:22:53.722 INFO:teuthology.orchestra.run.vm09.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T10:22:53.722 INFO:teuthology.orchestra.run.vm09.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T10:22:53.722 INFO:teuthology.orchestra.run.vm09.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:53.722 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:53.723 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:53.723 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:53.723 INFO:teuthology.orchestra.run.vm09.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:53.723 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:53.723 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:53.723 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:53.723 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:53.723 INFO:teuthology.orchestra.run.vm09.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:53.723 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:53.723 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:53.723 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T10:22:53.723 INFO:teuthology.orchestra.run.vm09.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T10:22:53.723 INFO:teuthology.orchestra.run.vm09.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:53.747 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:53.747 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:53.749 DEBUG:teuthology.orchestra.run.vm09:> dpkg -l | grep '^.\(U\|H\)R' | awk '{print $2}' | sudo xargs --no-run-if-empty dpkg -P --force-remove-reinstreq 2026-03-10T10:22:53.763 INFO:teuthology.orchestra.run.vm00.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117336 files and directories currently installed.) 2026-03-10T10:22:53.765 INFO:teuthology.orchestra.run.vm00.stdout:Removing ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:53.782 INFO:teuthology.orchestra.run.vm00.stdout:Removing jq (1.6-2.1ubuntu3.1) ... 2026-03-10T10:22:53.794 INFO:teuthology.orchestra.run.vm00.stdout:Removing kpartx (0.8.8-1ubuntu1.22.04.4) ... 2026-03-10T10:22:53.803 DEBUG:teuthology.orchestra.run.vm09:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" autoremove 2026-03-10T10:22:53.806 INFO:teuthology.orchestra.run.vm00.stdout:Removing libboost-iostreams1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-10T10:22:53.816 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:53.817 INFO:teuthology.orchestra.run.vm00.stdout:Removing libboost-thread1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-10T10:22:53.817 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:53.829 INFO:teuthology.orchestra.run.vm00.stdout:Removing libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-10T10:22:53.839 INFO:teuthology.orchestra.run.vm00.stdout:Removing libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:22:53.850 INFO:teuthology.orchestra.run.vm00.stdout:Removing libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:22:53.862 INFO:teuthology.orchestra.run.vm00.stdout:Removing libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:22:53.880 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:22:53.882 INFO:teuthology.orchestra.run.vm00.stdout:Removing libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-10T10:22:53.894 INFO:teuthology.orchestra.run.vm00.stdout:Removing libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-10T10:22:53.906 INFO:teuthology.orchestra.run.vm00.stdout:Removing libgfapi0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T10:22:53.917 INFO:teuthology.orchestra.run.vm00.stdout:Removing libgfrpc0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T10:22:53.932 INFO:teuthology.orchestra.run.vm00.stdout:Removing libgfxdr0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout:Package 'rbd-fuse' is not installed, so not removed 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout:The following packages were automatically installed and are no longer required: 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T10:22:53.933 INFO:teuthology.orchestra.run.vm03.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T10:22:53.934 INFO:teuthology.orchestra.run.vm03.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-10T10:22:53.943 INFO:teuthology.orchestra.run.vm00.stdout:Removing libglusterfs0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T10:22:53.947 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 0 to remove and 10 not upgraded. 2026-03-10T10:22:53.947 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:22:53.949 DEBUG:teuthology.orchestra.run.vm03:> dpkg -l | grep '^.\(U\|H\)R' | awk '{print $2}' | sudo xargs --no-run-if-empty dpkg -P --force-remove-reinstreq 2026-03-10T10:22:53.955 INFO:teuthology.orchestra.run.vm00.stdout:Removing libiscsi7:amd64 (1.19.0-3build2) ... 2026-03-10T10:22:53.966 INFO:teuthology.orchestra.run.vm00.stdout:Removing libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-10T10:22:53.977 INFO:teuthology.orchestra.run.vm00.stdout:Removing liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-10T10:22:53.988 INFO:teuthology.orchestra.run.vm00.stdout:Removing luarocks (3.8.0+dfsg1-1) ... 2026-03-10T10:22:54.002 DEBUG:teuthology.orchestra.run.vm03:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" autoremove 2026-03-10T10:22:54.014 INFO:teuthology.orchestra.run.vm00.stdout:Removing liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-10T10:22:54.024 INFO:teuthology.orchestra.run.vm00.stdout:Removing libnbd0 (1.10.5-1) ... 2026-03-10T10:22:54.034 INFO:teuthology.orchestra.run.vm00.stdout:Removing liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-10T10:22:54.043 INFO:teuthology.orchestra.run.vm00.stdout:Removing libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-10T10:22:54.051 INFO:teuthology.orchestra.run.vm00.stdout:Removing libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-10T10:22:54.061 INFO:teuthology.orchestra.run.vm00.stdout:Removing libpmemobj1:amd64 (1.11.1-3build1) ... 2026-03-10T10:22:54.071 INFO:teuthology.orchestra.run.vm00.stdout:Removing librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-10T10:22:54.080 INFO:teuthology.orchestra.run.vm09.stdout:Building dependency tree... 2026-03-10T10:22:54.081 INFO:teuthology.orchestra.run.vm09.stdout:Reading state information... 2026-03-10T10:22:54.081 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:22:54.083 INFO:teuthology.orchestra.run.vm00.stdout:Removing libreadline-dev:amd64 (8.1.2-1) ... 2026-03-10T10:22:54.094 INFO:teuthology.orchestra.run.vm00.stdout:Removing sg3-utils-udev (1.46-1ubuntu0.22.04.1) ... 2026-03-10T10:22:54.102 INFO:teuthology.orchestra.run.vm00.stdout:update-initramfs: deferring update (trigger activated) 2026-03-10T10:22:54.113 INFO:teuthology.orchestra.run.vm00.stdout:Removing sg3-utils (1.46-1ubuntu0.22.04.1) ... 2026-03-10T10:22:54.130 INFO:teuthology.orchestra.run.vm00.stdout:Removing libsgutils2-2:amd64 (1.46-1ubuntu0.22.04.1) ... 2026-03-10T10:22:54.142 INFO:teuthology.orchestra.run.vm00.stdout:Removing lua-any (27ubuntu1) ... 2026-03-10T10:22:54.154 INFO:teuthology.orchestra.run.vm00.stdout:Removing lua-sec:amd64 (1.0.2-1) ... 2026-03-10T10:22:54.167 INFO:teuthology.orchestra.run.vm00.stdout:Removing lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-10T10:22:54.182 INFO:teuthology.orchestra.run.vm00.stdout:Removing lua5.1 (5.1.5-8.1build4) ... 2026-03-10T10:22:54.200 INFO:teuthology.orchestra.run.vm00.stdout:Removing nvme-cli (1.16-3ubuntu0.3) ... 2026-03-10T10:22:54.239 INFO:teuthology.orchestra.run.vm09.stdout:The following packages will be REMOVED: 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:54.240 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T10:22:54.241 INFO:teuthology.orchestra.run.vm09.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T10:22:54.277 INFO:teuthology.orchestra.run.vm03.stdout:Building dependency tree... 2026-03-10T10:22:54.278 INFO:teuthology.orchestra.run.vm03.stdout:Reading state information... 2026-03-10T10:22:54.409 INFO:teuthology.orchestra.run.vm09.stdout:0 upgraded, 0 newly installed, 87 to remove and 10 not upgraded. 2026-03-10T10:22:54.410 INFO:teuthology.orchestra.run.vm09.stdout:After this operation, 107 MB disk space will be freed. 2026-03-10T10:22:54.422 INFO:teuthology.orchestra.run.vm03.stdout:The following packages will be REMOVED: 2026-03-10T10:22:54.422 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core jq kpartx libboost-iostreams1.74.0 2026-03-10T10:22:54.422 INFO:teuthology.orchestra.run.vm03.stdout: libboost-thread1.74.0 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-10T10:22:54.422 INFO:teuthology.orchestra.run.vm03.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 liblua5.3-dev libnbd0 2026-03-10T10:22:54.422 INFO:teuthology.orchestra.run.vm03.stdout: liboath0 libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 2026-03-10T10:22:54.422 INFO:teuthology.orchestra.run.vm03.stdout: libqt5network5 librdkafka1 libreadline-dev libsgutils2-2 libthrift-0.16.0 2026-03-10T10:22:54.423 INFO:teuthology.orchestra.run.vm03.stdout: lua-any lua-sec lua-socket lua5.1 luarocks nvme-cli pkg-config 2026-03-10T10:22:54.423 INFO:teuthology.orchestra.run.vm03.stdout: python-asyncssh-doc python-pastedeploy-tpl python3-asyncssh 2026-03-10T10:22:54.423 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-10T10:22:54.423 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-10T10:22:54.423 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-10T10:22:54.423 INFO:teuthology.orchestra.run.vm03.stdout: python3-joblib python3-kubernetes python3-logutils python3-mako 2026-03-10T10:22:54.423 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort python3-paste python3-pastedeploy python3-pastescript 2026-03-10T10:22:54.423 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan python3-portend python3-prettytable python3-psutil 2026-03-10T10:22:54.423 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyinotify python3-repoze.lru python3-requests-oauthlib 2026-03-10T10:22:54.423 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes python3-rsa python3-simplegeneric python3-simplejson 2026-03-10T10:22:54.423 INFO:teuthology.orchestra.run.vm03.stdout: python3-singledispatch python3-sklearn python3-sklearn-lib python3-tempita 2026-03-10T10:22:54.423 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora python3-threadpoolctl python3-waitress python3-wcwidth 2026-03-10T10:22:54.423 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob python3-websocket python3-webtest python3-werkzeug 2026-03-10T10:22:54.423 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc.lockfile qttranslations5-l10n sg3-utils sg3-utils-udev 2026-03-10T10:22:54.423 INFO:teuthology.orchestra.run.vm03.stdout: smartmontools socat unzip xmlstarlet zip 2026-03-10T10:22:54.446 INFO:teuthology.orchestra.run.vm09.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117336 files and directories currently installed.) 2026-03-10T10:22:54.448 INFO:teuthology.orchestra.run.vm09.stdout:Removing ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:54.466 INFO:teuthology.orchestra.run.vm09.stdout:Removing jq (1.6-2.1ubuntu3.1) ... 2026-03-10T10:22:54.478 INFO:teuthology.orchestra.run.vm09.stdout:Removing kpartx (0.8.8-1ubuntu1.22.04.4) ... 2026-03-10T10:22:54.491 INFO:teuthology.orchestra.run.vm09.stdout:Removing libboost-iostreams1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-10T10:22:54.504 INFO:teuthology.orchestra.run.vm09.stdout:Removing libboost-thread1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-10T10:22:54.516 INFO:teuthology.orchestra.run.vm09.stdout:Removing libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-10T10:22:54.527 INFO:teuthology.orchestra.run.vm09.stdout:Removing libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:22:54.538 INFO:teuthology.orchestra.run.vm09.stdout:Removing libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:22:54.548 INFO:teuthology.orchestra.run.vm09.stdout:Removing libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:22:54.566 INFO:teuthology.orchestra.run.vm09.stdout:Removing libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-10T10:22:54.577 INFO:teuthology.orchestra.run.vm09.stdout:Removing libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-10T10:22:54.590 INFO:teuthology.orchestra.run.vm09.stdout:Removing libgfapi0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T10:22:54.593 INFO:teuthology.orchestra.run.vm03.stdout:0 upgraded, 0 newly installed, 87 to remove and 10 not upgraded. 2026-03-10T10:22:54.593 INFO:teuthology.orchestra.run.vm03.stdout:After this operation, 107 MB disk space will be freed. 2026-03-10T10:22:54.602 INFO:teuthology.orchestra.run.vm09.stdout:Removing libgfrpc0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T10:22:54.613 INFO:teuthology.orchestra.run.vm09.stdout:Removing libgfxdr0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T10:22:54.620 INFO:teuthology.orchestra.run.vm00.stdout:Removing pkg-config (0.29.2-1ubuntu3) ... 2026-03-10T10:22:54.624 INFO:teuthology.orchestra.run.vm09.stdout:Removing libglusterfs0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T10:22:54.632 INFO:teuthology.orchestra.run.vm03.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 117336 files and directories currently installed.) 2026-03-10T10:22:54.633 INFO:teuthology.orchestra.run.vm03.stdout:Removing ceph-mgr-modules-core (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:54.636 INFO:teuthology.orchestra.run.vm09.stdout:Removing libiscsi7:amd64 (1.19.0-3build2) ... 2026-03-10T10:22:54.647 INFO:teuthology.orchestra.run.vm09.stdout:Removing libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-10T10:22:54.650 INFO:teuthology.orchestra.run.vm03.stdout:Removing jq (1.6-2.1ubuntu3.1) ... 2026-03-10T10:22:54.653 INFO:teuthology.orchestra.run.vm00.stdout:Removing python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-10T10:22:54.658 INFO:teuthology.orchestra.run.vm09.stdout:Removing liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-10T10:22:54.661 INFO:teuthology.orchestra.run.vm03.stdout:Removing kpartx (0.8.8-1ubuntu1.22.04.4) ... 2026-03-10T10:22:54.672 INFO:teuthology.orchestra.run.vm09.stdout:Removing luarocks (3.8.0+dfsg1-1) ... 2026-03-10T10:22:54.674 INFO:teuthology.orchestra.run.vm03.stdout:Removing libboost-iostreams1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-10T10:22:54.726 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-pecan (1.3.3-4ubuntu2) ... 2026-03-10T10:22:54.732 INFO:teuthology.orchestra.run.vm03.stdout:Removing libboost-thread1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-10T10:22:54.735 INFO:teuthology.orchestra.run.vm09.stdout:Removing liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-10T10:22:54.742 INFO:teuthology.orchestra.run.vm03.stdout:Removing libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-10T10:22:54.747 INFO:teuthology.orchestra.run.vm09.stdout:Removing libnbd0 (1.10.5-1) ... 2026-03-10T10:22:54.752 INFO:teuthology.orchestra.run.vm03.stdout:Removing libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:22:54.759 INFO:teuthology.orchestra.run.vm09.stdout:Removing liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-10T10:22:54.761 INFO:teuthology.orchestra.run.vm03.stdout:Removing libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:22:54.771 INFO:teuthology.orchestra.run.vm09.stdout:Removing libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-10T10:22:54.771 INFO:teuthology.orchestra.run.vm03.stdout:Removing libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-10T10:22:54.783 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-webtest (2.0.35-1) ... 2026-03-10T10:22:54.783 INFO:teuthology.orchestra.run.vm09.stdout:Removing libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-10T10:22:54.788 INFO:teuthology.orchestra.run.vm03.stdout:Removing libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-10T10:22:54.796 INFO:teuthology.orchestra.run.vm09.stdout:Removing libpmemobj1:amd64 (1.11.1-3build1) ... 2026-03-10T10:22:54.798 INFO:teuthology.orchestra.run.vm03.stdout:Removing libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-10T10:22:54.807 INFO:teuthology.orchestra.run.vm03.stdout:Removing libgfapi0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T10:22:54.808 INFO:teuthology.orchestra.run.vm09.stdout:Removing librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-10T10:22:54.817 INFO:teuthology.orchestra.run.vm03.stdout:Removing libgfrpc0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T10:22:54.820 INFO:teuthology.orchestra.run.vm09.stdout:Removing libreadline-dev:amd64 (8.1.2-1) ... 2026-03-10T10:22:54.827 INFO:teuthology.orchestra.run.vm03.stdout:Removing libgfxdr0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T10:22:54.831 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-pastescript (2.0.2-4) ... 2026-03-10T10:22:54.832 INFO:teuthology.orchestra.run.vm09.stdout:Removing sg3-utils-udev (1.46-1ubuntu0.22.04.1) ... 2026-03-10T10:22:54.836 INFO:teuthology.orchestra.run.vm03.stdout:Removing libglusterfs0:amd64 (10.1-1ubuntu0.2) ... 2026-03-10T10:22:54.840 INFO:teuthology.orchestra.run.vm09.stdout:update-initramfs: deferring update (trigger activated) 2026-03-10T10:22:54.848 INFO:teuthology.orchestra.run.vm03.stdout:Removing libiscsi7:amd64 (1.19.0-3build2) ... 2026-03-10T10:22:54.851 INFO:teuthology.orchestra.run.vm09.stdout:Removing sg3-utils (1.46-1ubuntu0.22.04.1) ... 2026-03-10T10:22:54.859 INFO:teuthology.orchestra.run.vm03.stdout:Removing libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-10T10:22:54.869 INFO:teuthology.orchestra.run.vm09.stdout:Removing libsgutils2-2:amd64 (1.46-1ubuntu0.22.04.1) ... 2026-03-10T10:22:54.870 INFO:teuthology.orchestra.run.vm03.stdout:Removing liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-10T10:22:54.882 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-pastedeploy (2.1.1-1) ... 2026-03-10T10:22:54.882 INFO:teuthology.orchestra.run.vm03.stdout:Removing luarocks (3.8.0+dfsg1-1) ... 2026-03-10T10:22:54.882 INFO:teuthology.orchestra.run.vm09.stdout:Removing lua-any (27ubuntu1) ... 2026-03-10T10:22:54.893 INFO:teuthology.orchestra.run.vm09.stdout:Removing lua-sec:amd64 (1.0.2-1) ... 2026-03-10T10:22:54.907 INFO:teuthology.orchestra.run.vm09.stdout:Removing lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-10T10:22:54.907 INFO:teuthology.orchestra.run.vm03.stdout:Removing liblua5.3-dev:amd64 (5.3.6-1build1) ... 2026-03-10T10:22:54.918 INFO:teuthology.orchestra.run.vm03.stdout:Removing libnbd0 (1.10.5-1) ... 2026-03-10T10:22:54.921 INFO:teuthology.orchestra.run.vm09.stdout:Removing lua5.1 (5.1.5-8.1build4) ... 2026-03-10T10:22:54.928 INFO:teuthology.orchestra.run.vm03.stdout:Removing liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-10T10:22:54.931 INFO:teuthology.orchestra.run.vm00.stdout:Removing python-pastedeploy-tpl (2.1.1-1) ... 2026-03-10T10:22:54.937 INFO:teuthology.orchestra.run.vm03.stdout:Removing libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-10T10:22:54.938 INFO:teuthology.orchestra.run.vm09.stdout:Removing nvme-cli (1.16-3ubuntu0.3) ... 2026-03-10T10:22:54.944 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-10T10:22:54.946 INFO:teuthology.orchestra.run.vm03.stdout:Removing libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-10T10:22:54.955 INFO:teuthology.orchestra.run.vm03.stdout:Removing libpmemobj1:amd64 (1.11.1-3build1) ... 2026-03-10T10:22:54.966 INFO:teuthology.orchestra.run.vm03.stdout:Removing librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-10T10:22:54.977 INFO:teuthology.orchestra.run.vm03.stdout:Removing libreadline-dev:amd64 (8.1.2-1) ... 2026-03-10T10:22:54.988 INFO:teuthology.orchestra.run.vm03.stdout:Removing sg3-utils-udev (1.46-1ubuntu0.22.04.1) ... 2026-03-10T10:22:54.995 INFO:teuthology.orchestra.run.vm03.stdout:update-initramfs: deferring update (trigger activated) 2026-03-10T10:22:55.004 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-10T10:22:55.006 INFO:teuthology.orchestra.run.vm03.stdout:Removing sg3-utils (1.46-1ubuntu0.22.04.1) ... 2026-03-10T10:22:55.023 INFO:teuthology.orchestra.run.vm03.stdout:Removing libsgutils2-2:amd64 (1.46-1ubuntu0.22.04.1) ... 2026-03-10T10:22:55.034 INFO:teuthology.orchestra.run.vm03.stdout:Removing lua-any (27ubuntu1) ... 2026-03-10T10:22:55.046 INFO:teuthology.orchestra.run.vm03.stdout:Removing lua-sec:amd64 (1.0.2-1) ... 2026-03-10T10:22:55.058 INFO:teuthology.orchestra.run.vm03.stdout:Removing lua-socket:amd64 (3.0~rc1+git+ac3201d-6) ... 2026-03-10T10:22:55.071 INFO:teuthology.orchestra.run.vm03.stdout:Removing lua5.1 (5.1.5-8.1build4) ... 2026-03-10T10:22:55.089 INFO:teuthology.orchestra.run.vm03.stdout:Removing nvme-cli (1.16-3ubuntu0.3) ... 2026-03-10T10:22:55.288 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-google-auth (1.5.1-3) ... 2026-03-10T10:22:55.342 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-cachetools (5.0.0-1) ... 2026-03-10T10:22:55.354 INFO:teuthology.orchestra.run.vm09.stdout:Removing pkg-config (0.29.2-1ubuntu3) ... 2026-03-10T10:22:55.385 INFO:teuthology.orchestra.run.vm09.stdout:Removing python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-10T10:22:55.391 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:55.416 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-pecan (1.3.3-4ubuntu2) ... 2026-03-10T10:22:55.441 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:55.476 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-webtest (2.0.35-1) ... 2026-03-10T10:22:55.484 INFO:teuthology.orchestra.run.vm03.stdout:Removing pkg-config (0.29.2-1ubuntu3) ... 2026-03-10T10:22:55.495 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-cherrypy3 (18.6.1-4) ... 2026-03-10T10:22:55.515 INFO:teuthology.orchestra.run.vm03.stdout:Removing python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-10T10:22:55.525 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-pastescript (2.0.2-4) ... 2026-03-10T10:22:55.541 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-pecan (1.3.3-4ubuntu2) ... 2026-03-10T10:22:55.558 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-10T10:22:55.575 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-pastedeploy (2.1.1-1) ... 2026-03-10T10:22:55.597 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-webtest (2.0.35-1) ... 2026-03-10T10:22:55.609 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-jaraco.collections (3.4.0-2) ... 2026-03-10T10:22:55.627 INFO:teuthology.orchestra.run.vm09.stdout:Removing python-pastedeploy-tpl (2.1.1-1) ... 2026-03-10T10:22:55.638 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-10T10:22:55.644 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-pastescript (2.0.2-4) ... 2026-03-10T10:22:55.657 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-jaraco.classes (3.2.1-3) ... 2026-03-10T10:22:55.694 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-10T10:22:55.697 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-pastedeploy (2.1.1-1) ... 2026-03-10T10:22:55.706 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-portend (3.0.0-1) ... 2026-03-10T10:22:55.747 INFO:teuthology.orchestra.run.vm03.stdout:Removing python-pastedeploy-tpl (2.1.1-1) ... 2026-03-10T10:22:55.753 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-tempora (4.1.2-1) ... 2026-03-10T10:22:55.757 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-10T10:22:55.801 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-jaraco.text (3.6.0-2) ... 2026-03-10T10:22:55.811 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-10T10:22:55.849 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-jaraco.functools (3.4.0-2) ... 2026-03-10T10:22:55.901 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-10T10:22:55.964 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-google-auth (1.5.1-3) ... 2026-03-10T10:22:56.017 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-cachetools (5.0.0-1) ... 2026-03-10T10:22:56.021 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-joblib (0.17.0-4ubuntu1) ... 2026-03-10T10:22:56.066 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:56.079 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-google-auth (1.5.1-3) ... 2026-03-10T10:22:56.084 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-logutils (0.3.3-8) ... 2026-03-10T10:22:56.112 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:56.130 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-cachetools (5.0.0-1) ... 2026-03-10T10:22:56.132 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-10T10:22:56.163 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-cherrypy3 (18.6.1-4) ... 2026-03-10T10:22:56.178 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-ceph-argparse (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:56.184 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-natsort (8.0.2-1) ... 2026-03-10T10:22:56.222 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-10T10:22:56.226 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-ceph-common (19.2.3-678-ge911bdeb-1jammy) ... 2026-03-10T10:22:56.233 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-paste (3.5.0+dfsg1-1) ... 2026-03-10T10:22:56.273 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-jaraco.collections (3.4.0-2) ... 2026-03-10T10:22:56.276 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-cherrypy3 (18.6.1-4) ... 2026-03-10T10:22:56.291 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-prettytable (2.5.0-2) ... 2026-03-10T10:22:56.321 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-jaraco.classes (3.2.1-3) ... 2026-03-10T10:22:56.333 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-10T10:22:56.337 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-psutil (5.9.0-1build1) ... 2026-03-10T10:22:56.369 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-portend (3.0.0-1) ... 2026-03-10T10:22:56.381 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-jaraco.collections (3.4.0-2) ... 2026-03-10T10:22:56.389 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-pyinotify (0.9.6-1.3) ... 2026-03-10T10:22:56.421 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-tempora (4.1.2-1) ... 2026-03-10T10:22:56.424 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-jaraco.classes (3.2.1-3) ... 2026-03-10T10:22:56.437 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-routes (2.5.1-1ubuntu1) ... 2026-03-10T10:22:56.469 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-jaraco.text (3.6.0-2) ... 2026-03-10T10:22:56.469 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-portend (3.0.0-1) ... 2026-03-10T10:22:56.490 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-repoze.lru (0.7-2) ... 2026-03-10T10:22:56.517 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-jaraco.functools (3.4.0-2) ... 2026-03-10T10:22:56.519 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-tempora (4.1.2-1) ... 2026-03-10T10:22:56.539 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-10T10:22:56.565 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-10T10:22:56.565 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-jaraco.text (3.6.0-2) ... 2026-03-10T10:22:56.587 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-rsa (4.8-1) ... 2026-03-10T10:22:56.612 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-jaraco.functools (3.4.0-2) ... 2026-03-10T10:22:56.637 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-simplegeneric (0.8.1-3) ... 2026-03-10T10:22:56.659 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-10T10:22:56.685 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-simplejson (3.17.6-1build1) ... 2026-03-10T10:22:56.687 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-joblib (0.17.0-4ubuntu1) ... 2026-03-10T10:22:56.738 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-singledispatch (3.4.0.3-3) ... 2026-03-10T10:22:56.749 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-logutils (0.3.3-8) ... 2026-03-10T10:22:56.783 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-joblib (0.17.0-4ubuntu1) ... 2026-03-10T10:22:56.785 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-10T10:22:56.798 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-10T10:22:56.810 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-tempita (0.5.2-6ubuntu1) ... 2026-03-10T10:22:56.841 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-logutils (0.3.3-8) ... 2026-03-10T10:22:56.849 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-natsort (8.0.2-1) ... 2026-03-10T10:22:56.857 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-threadpoolctl (3.1.0-1) ... 2026-03-10T10:22:56.891 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-mako (1.1.3+ds1-2ubuntu0.1) ... 2026-03-10T10:22:56.899 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-paste (3.5.0+dfsg1-1) ... 2026-03-10T10:22:56.903 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-10T10:22:56.941 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-natsort (8.0.2-1) ... 2026-03-10T10:22:56.953 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-10T10:22:56.959 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-prettytable (2.5.0-2) ... 2026-03-10T10:22:56.990 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-paste (3.5.0+dfsg1-1) ... 2026-03-10T10:22:57.003 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-10T10:22:57.007 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-psutil (5.9.0-1build1) ... 2026-03-10T10:22:57.047 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-prettytable (2.5.0-2) ... 2026-03-10T10:22:57.052 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-websocket (1.2.3-1) ... 2026-03-10T10:22:57.056 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-pyinotify (0.9.6-1.3) ... 2026-03-10T10:22:57.093 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-psutil (5.9.0-1build1) ... 2026-03-10T10:22:57.102 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-10T10:22:57.102 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-routes (2.5.1-1ubuntu1) ... 2026-03-10T10:22:57.144 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-pyinotify (0.9.6-1.3) ... 2026-03-10T10:22:57.153 INFO:teuthology.orchestra.run.vm00.stdout:Removing python3-zc.lockfile (2.0-1) ... 2026-03-10T10:22:57.155 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-repoze.lru (0.7-2) ... 2026-03-10T10:22:57.191 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-routes (2.5.1-1ubuntu1) ... 2026-03-10T10:22:57.200 INFO:teuthology.orchestra.run.vm00.stdout:Removing qttranslations5-l10n (5.15.3-1) ... 2026-03-10T10:22:57.204 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-10T10:22:57.223 INFO:teuthology.orchestra.run.vm00.stdout:Removing smartmontools (7.2-1ubuntu0.1) ... 2026-03-10T10:22:57.241 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-repoze.lru (0.7-2) ... 2026-03-10T10:22:57.253 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-rsa (4.8-1) ... 2026-03-10T10:22:57.287 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-10T10:22:57.303 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-simplegeneric (0.8.1-3) ... 2026-03-10T10:22:57.335 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-rsa (4.8-1) ... 2026-03-10T10:22:57.349 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-simplejson (3.17.6-1build1) ... 2026-03-10T10:22:57.387 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-simplegeneric (0.8.1-3) ... 2026-03-10T10:22:57.401 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-singledispatch (3.4.0.3-3) ... 2026-03-10T10:22:57.433 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-simplejson (3.17.6-1build1) ... 2026-03-10T10:22:57.448 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-10T10:22:57.475 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-tempita (0.5.2-6ubuntu1) ... 2026-03-10T10:22:57.488 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-singledispatch (3.4.0.3-3) ... 2026-03-10T10:22:57.521 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-threadpoolctl (3.1.0-1) ... 2026-03-10T10:22:57.536 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-10T10:22:57.560 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-tempita (0.5.2-6ubuntu1) ... 2026-03-10T10:22:57.566 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-10T10:22:57.608 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-threadpoolctl (3.1.0-1) ... 2026-03-10T10:22:57.614 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-10T10:22:57.619 INFO:teuthology.orchestra.run.vm00.stdout:Removing socat (1.7.4.1-3ubuntu4) ... 2026-03-10T10:22:57.632 INFO:teuthology.orchestra.run.vm00.stdout:Removing unzip (6.0-26ubuntu3.2) ... 2026-03-10T10:22:57.651 INFO:teuthology.orchestra.run.vm00.stdout:Removing xmlstarlet (1.6.1-2.1) ... 2026-03-10T10:22:57.656 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-waitress (1.4.4-1.1ubuntu1.1) ... 2026-03-10T10:22:57.661 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-10T10:22:57.668 INFO:teuthology.orchestra.run.vm00.stdout:Removing zip (3.0-12build2) ... 2026-03-10T10:22:57.694 INFO:teuthology.orchestra.run.vm00.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T10:22:57.704 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-10T10:22:57.705 INFO:teuthology.orchestra.run.vm00.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:57.707 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-websocket (1.2.3-1) ... 2026-03-10T10:22:57.748 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-10T10:22:57.753 INFO:teuthology.orchestra.run.vm00.stdout:Processing triggers for mailcap (3.70+nmu1ubuntu1) ... 2026-03-10T10:22:57.756 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-10T10:22:57.760 INFO:teuthology.orchestra.run.vm00.stdout:Processing triggers for initramfs-tools (0.140ubuntu13.5) ... 2026-03-10T10:22:57.776 INFO:teuthology.orchestra.run.vm00.stdout:update-initramfs: Generating /boot/initrd.img-5.15.0-1092-kvm 2026-03-10T10:22:57.795 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-websocket (1.2.3-1) ... 2026-03-10T10:22:57.806 INFO:teuthology.orchestra.run.vm09.stdout:Removing python3-zc.lockfile (2.0-1) ... 2026-03-10T10:22:57.840 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-werkzeug (2.0.2+dfsg1-1ubuntu0.22.04.3) ... 2026-03-10T10:22:57.851 INFO:teuthology.orchestra.run.vm09.stdout:Removing qttranslations5-l10n (5.15.3-1) ... 2026-03-10T10:22:57.873 INFO:teuthology.orchestra.run.vm09.stdout:Removing smartmontools (7.2-1ubuntu0.1) ... 2026-03-10T10:22:57.889 INFO:teuthology.orchestra.run.vm03.stdout:Removing python3-zc.lockfile (2.0-1) ... 2026-03-10T10:22:57.936 INFO:teuthology.orchestra.run.vm03.stdout:Removing qttranslations5-l10n (5.15.3-1) ... 2026-03-10T10:22:57.957 INFO:teuthology.orchestra.run.vm03.stdout:Removing smartmontools (7.2-1ubuntu0.1) ... 2026-03-10T10:22:58.294 INFO:teuthology.orchestra.run.vm09.stdout:Removing socat (1.7.4.1-3ubuntu4) ... 2026-03-10T10:22:58.306 INFO:teuthology.orchestra.run.vm09.stdout:Removing unzip (6.0-26ubuntu3.2) ... 2026-03-10T10:22:58.326 INFO:teuthology.orchestra.run.vm09.stdout:Removing xmlstarlet (1.6.1-2.1) ... 2026-03-10T10:22:58.345 INFO:teuthology.orchestra.run.vm09.stdout:Removing zip (3.0-12build2) ... 2026-03-10T10:22:58.357 INFO:teuthology.orchestra.run.vm03.stdout:Removing socat (1.7.4.1-3ubuntu4) ... 2026-03-10T10:22:58.369 INFO:teuthology.orchestra.run.vm03.stdout:Removing unzip (6.0-26ubuntu3.2) ... 2026-03-10T10:22:58.370 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T10:22:58.379 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:58.388 INFO:teuthology.orchestra.run.vm03.stdout:Removing xmlstarlet (1.6.1-2.1) ... 2026-03-10T10:22:58.406 INFO:teuthology.orchestra.run.vm03.stdout:Removing zip (3.0-12build2) ... 2026-03-10T10:22:58.421 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for mailcap (3.70+nmu1ubuntu1) ... 2026-03-10T10:22:58.429 INFO:teuthology.orchestra.run.vm09.stdout:Processing triggers for initramfs-tools (0.140ubuntu13.5) ... 2026-03-10T10:22:58.430 INFO:teuthology.orchestra.run.vm03.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-10T10:22:58.440 INFO:teuthology.orchestra.run.vm03.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-10T10:22:58.445 INFO:teuthology.orchestra.run.vm09.stdout:update-initramfs: Generating /boot/initrd.img-5.15.0-1092-kvm 2026-03-10T10:22:58.482 INFO:teuthology.orchestra.run.vm03.stdout:Processing triggers for mailcap (3.70+nmu1ubuntu1) ... 2026-03-10T10:22:58.490 INFO:teuthology.orchestra.run.vm03.stdout:Processing triggers for initramfs-tools (0.140ubuntu13.5) ... 2026-03-10T10:22:58.509 INFO:teuthology.orchestra.run.vm03.stdout:update-initramfs: Generating /boot/initrd.img-5.15.0-1092-kvm 2026-03-10T10:22:59.249 INFO:teuthology.orchestra.run.vm00.stdout:W: mkconf: MD subsystem is not loaded, thus I cannot scan for arrays. 2026-03-10T10:22:59.249 INFO:teuthology.orchestra.run.vm00.stdout:W: mdadm: failed to auto-generate temporary mdadm.conf file. 2026-03-10T10:22:59.923 INFO:teuthology.orchestra.run.vm09.stdout:W: mkconf: MD subsystem is not loaded, thus I cannot scan for arrays. 2026-03-10T10:22:59.924 INFO:teuthology.orchestra.run.vm09.stdout:W: mdadm: failed to auto-generate temporary mdadm.conf file. 2026-03-10T10:22:59.965 INFO:teuthology.orchestra.run.vm03.stdout:W: mkconf: MD subsystem is not loaded, thus I cannot scan for arrays. 2026-03-10T10:22:59.965 INFO:teuthology.orchestra.run.vm03.stdout:W: mdadm: failed to auto-generate temporary mdadm.conf file. 2026-03-10T10:23:01.077 INFO:teuthology.orchestra.run.vm00.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:23:01.100 DEBUG:teuthology.parallel:result is None 2026-03-10T10:23:02.020 INFO:teuthology.orchestra.run.vm03.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:23:02.022 DEBUG:teuthology.parallel:result is None 2026-03-10T10:23:02.057 INFO:teuthology.orchestra.run.vm09.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-10T10:23:02.059 DEBUG:teuthology.parallel:result is None 2026-03-10T10:23:02.059 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm00.local 2026-03-10T10:23:02.059 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm03.local 2026-03-10T10:23:02.059 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm09.local 2026-03-10T10:23:02.059 DEBUG:teuthology.orchestra.run.vm00:> sudo rm -f /etc/apt/sources.list.d/ceph.list 2026-03-10T10:23:02.059 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f /etc/apt/sources.list.d/ceph.list 2026-03-10T10:23:02.059 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/apt/sources.list.d/ceph.list 2026-03-10T10:23:02.068 DEBUG:teuthology.orchestra.run.vm00:> sudo apt-get update 2026-03-10T10:23:02.074 DEBUG:teuthology.orchestra.run.vm03:> sudo apt-get update 2026-03-10T10:23:02.110 DEBUG:teuthology.orchestra.run.vm09:> sudo apt-get update 2026-03-10T10:23:02.398 INFO:teuthology.orchestra.run.vm09.stdout:Hit:1 https://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-10T10:23:02.430 INFO:teuthology.orchestra.run.vm09.stdout:Hit:2 https://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-10T10:23:02.455 INFO:teuthology.orchestra.run.vm09.stdout:Hit:3 https://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-10T10:23:02.466 INFO:teuthology.orchestra.run.vm09.stdout:Hit:4 https://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-10T10:23:02.598 INFO:teuthology.orchestra.run.vm03.stdout:Hit:1 https://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-10T10:23:02.599 INFO:teuthology.orchestra.run.vm00.stdout:Hit:1 https://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-10T10:23:02.619 INFO:teuthology.orchestra.run.vm00.stdout:Hit:2 https://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-10T10:23:02.643 INFO:teuthology.orchestra.run.vm03.stdout:Hit:2 https://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-10T10:23:02.723 INFO:teuthology.orchestra.run.vm00.stdout:Hit:3 https://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-10T10:23:02.755 INFO:teuthology.orchestra.run.vm03.stdout:Hit:3 https://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-10T10:23:02.827 INFO:teuthology.orchestra.run.vm00.stdout:Hit:4 https://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-10T10:23:02.867 INFO:teuthology.orchestra.run.vm03.stdout:Hit:4 https://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-10T10:23:03.335 INFO:teuthology.orchestra.run.vm09.stdout:Reading package lists... 2026-03-10T10:23:03.349 DEBUG:teuthology.parallel:result is None 2026-03-10T10:23:03.631 INFO:teuthology.orchestra.run.vm00.stdout:Reading package lists... 2026-03-10T10:23:03.644 DEBUG:teuthology.parallel:result is None 2026-03-10T10:23:03.714 INFO:teuthology.orchestra.run.vm03.stdout:Reading package lists... 2026-03-10T10:23:03.725 DEBUG:teuthology.parallel:result is None 2026-03-10T10:23:03.725 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-10T10:23:03.728 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-10T10:23:03.728 DEBUG:teuthology.orchestra.run.vm00:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T10:23:03.729 DEBUG:teuthology.orchestra.run.vm03:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T10:23:03.730 DEBUG:teuthology.orchestra.run.vm09:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T10:23:03.913 INFO:teuthology.orchestra.run.vm03.stdout: remote refid st t when poll reach delay offset jitter 2026-03-10T10:23:03.913 INFO:teuthology.orchestra.run.vm03.stdout:============================================================================== 2026-03-10T10:23:03.913 INFO:teuthology.orchestra.run.vm03.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:23:03.914 INFO:teuthology.orchestra.run.vm03.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:23:03.914 INFO:teuthology.orchestra.run.vm03.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:23:03.914 INFO:teuthology.orchestra.run.vm03.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:23:03.914 INFO:teuthology.orchestra.run.vm03.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:23:03.914 INFO:teuthology.orchestra.run.vm03.stdout:-ctb01.martinmoe 87.63.200.138 2 u 48 64 377 31.490 +0.325 0.157 2026-03-10T10:23:03.914 INFO:teuthology.orchestra.run.vm03.stdout:+mail.fu110.de 35.73.197.144 2 u 44 64 377 25.084 +0.331 0.232 2026-03-10T10:23:03.914 INFO:teuthology.orchestra.run.vm03.stdout:+static.222.16.4 35.73.197.144 2 u 40 64 377 0.720 +0.063 1.442 2026-03-10T10:23:03.914 INFO:teuthology.orchestra.run.vm03.stdout:*79.133.44.142 .MBGh. 1 u 41 64 377 20.481 -0.022 0.090 2026-03-10T10:23:03.914 INFO:teuthology.orchestra.run.vm03.stdout:-ntp1.lwlcom.net .GPS. 1 u 112 128 377 30.822 +3.320 0.093 2026-03-10T10:23:03.914 INFO:teuthology.orchestra.run.vm03.stdout:-where-you.at 31.209.85.243 2 u 46 64 377 24.938 +0.067 0.040 2026-03-10T10:23:03.914 INFO:teuthology.orchestra.run.vm03.stdout:+zeus.f5s.de 195.145.119.188 2 u 47 64 377 25.042 -0.012 0.045 2026-03-10T10:23:03.914 INFO:teuthology.orchestra.run.vm03.stdout:-vps-nue1.orlean 195.145.119.188 2 u 44 64 377 28.244 -1.980 0.297 2026-03-10T10:23:03.914 INFO:teuthology.orchestra.run.vm03.stdout:-vps-fra1.orlean 195.145.119.188 2 u 104 128 377 22.011 +0.077 6.200 2026-03-10T10:23:03.914 INFO:teuthology.orchestra.run.vm03.stdout:+185.252.140.125 216.239.35.4 2 u 42 64 377 25.111 +0.050 0.049 2026-03-10T10:23:04.078 INFO:teuthology.orchestra.run.vm09.stdout: remote refid st t when poll reach delay offset jitter 2026-03-10T10:23:04.078 INFO:teuthology.orchestra.run.vm09.stdout:============================================================================== 2026-03-10T10:23:04.078 INFO:teuthology.orchestra.run.vm09.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout:+ntp1.lwlcom.net .GPS. 1 u 47 64 377 30.825 +3.545 2.271 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout:+zeus.f5s.de 195.145.119.188 2 u 45 64 377 25.002 +0.172 2.253 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout:+where-you.at 31.209.85.243 2 u 46 64 377 24.938 -3.030 3.307 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout:+mail.fu110.de 35.73.197.144 2 u 43 64 377 25.051 +1.009 2.535 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout:+ctb01.martinmoe 87.63.200.138 2 u 53 64 377 31.603 +0.430 2.214 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout:+vsrv02141.custo 79.133.44.137 2 u 55 64 377 32.625 +0.646 2.432 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout:*79.133.44.142 .MBGh. 1 u 45 64 377 20.503 +1.645 3.219 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout:#139-162-152-20. 82.43.52.28 2 u 43 64 377 22.673 -8.218 3.455 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout:+mailout04.fisch 205.46.178.169 2 u 42 64 377 25.142 +0.427 2.277 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout:+stratum2-2.NTP. 129.70.137.82 2 u 11 64 37 30.348 -0.173 0.372 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout:+static.222.16.4 35.73.197.144 2 u 49 64 377 0.343 +0.401 2.239 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout:+185.252.140.125 216.239.35.4 2 u 42 64 377 25.102 +0.648 2.509 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout:+vps-nue1.orlean 195.145.119.188 2 u 52 64 377 28.302 -5.065 3.290 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout:+vps-fra1.orlean 195.145.119.188 2 u 39 64 377 21.976 -2.741 3.467 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout:#104-167-24-26.l 185.232.69.65 2 u 48 64 377 26.548 -1.890 2.769 2026-03-10T10:23:04.079 INFO:teuthology.orchestra.run.vm09.stdout:#alphyn.canonica 132.163.96.1 2 u 2 64 377 102.049 -2.344 1.963 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout: remote refid st t when poll reach delay offset jitter 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout:============================================================================== 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout:+ctb01.martinmoe 87.63.200.138 2 u 51 64 377 34.631 -2.239 1.678 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout:+static.222.16.4 35.73.197.144 2 u 53 64 377 0.351 +1.788 2.333 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout:+zeus.f5s.de 195.145.119.188 2 u 46 64 377 25.009 -0.942 1.726 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout:+mail.fu110.de 35.73.197.144 2 u 40 64 377 25.094 -0.332 1.568 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout:+ntp1.lwlcom.net .GPS. 1 u 43 64 377 30.846 +2.775 1.531 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout:+vsrv02141.custo 79.133.44.137 2 u 46 64 77 32.796 +1.846 1.953 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout:*79.133.44.142 .MBGh. 1 u 37 64 377 20.527 -0.214 1.420 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout:#139-162-152-20. 82.43.52.28 2 u 45 64 377 22.583 -3.251 2.289 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout:+mailout04.fisch 205.46.178.169 2 u 47 64 377 25.180 -0.054 1.398 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout:+where-you.at 31.209.85.243 2 u 45 64 377 24.983 -1.608 2.268 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout:+185.252.140.125 216.239.35.4 2 u 48 64 377 25.127 +1.683 2.255 2026-03-10T10:23:04.109 INFO:teuthology.orchestra.run.vm00.stdout:#185.125.190.56 79.243.60.50 2 u 64 64 377 35.291 -1.320 1.656 2026-03-10T10:23:04.109 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-10T10:23:04.112 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-10T10:23:04.112 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-10T10:23:04.114 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-10T10:23:04.131 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-10T10:23:04.133 INFO:teuthology.task.internal:Duration was 865.242201 seconds 2026-03-10T10:23:04.133 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-10T10:23:04.135 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-10T10:23:04.135 DEBUG:teuthology.orchestra.run.vm00:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-10T10:23:04.136 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-10T10:23:04.138 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-10T10:23:04.161 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-10T10:23:04.161 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm00.local 2026-03-10T10:23:04.161 DEBUG:teuthology.orchestra.run.vm00:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-10T10:23:04.216 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm03.local 2026-03-10T10:23:04.216 DEBUG:teuthology.orchestra.run.vm03:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-10T10:23:04.226 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm09.local 2026-03-10T10:23:04.227 DEBUG:teuthology.orchestra.run.vm09:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-10T10:23:04.238 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-10T10:23:04.238 DEBUG:teuthology.orchestra.run.vm00:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T10:23:04.260 DEBUG:teuthology.orchestra.run.vm03:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T10:23:04.271 DEBUG:teuthology.orchestra.run.vm09:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T10:23:04.321 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-10T10:23:04.321 DEBUG:teuthology.orchestra.run.vm00:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T10:23:04.344 DEBUG:teuthology.orchestra.run.vm03:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T10:23:04.351 DEBUG:teuthology.orchestra.run.vm09:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T10:23:04.351 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T10:23:04.351 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T10:23:04.351 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T10:23:04.352 INFO:teuthology.orchestra.run.vm00.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-10T10:23:04.352 INFO:teuthology.orchestra.run.vm00.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-10T10:23:04.356 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T10:23:04.357 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T10:23:04.357 INFO:teuthology.orchestra.run.vm03.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0%gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T10:23:04.357 INFO:teuthology.orchestra.run.vm03.stderr: -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-10T10:23:04.357 INFO:teuthology.orchestra.run.vm03.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-10T10:23:04.360 INFO:teuthology.orchestra.run.vm00.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 89.5% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-10T10:23:04.364 INFO:teuthology.orchestra.run.vm03.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 89.7% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-10T10:23:04.369 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T10:23:04.370 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T10:23:04.370 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T10:23:04.370 INFO:teuthology.orchestra.run.vm09.stderr: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-10T10:23:04.370 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-10T10:23:04.377 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 89.7% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-10T10:23:04.379 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-10T10:23:04.381 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-10T10:23:04.381 DEBUG:teuthology.orchestra.run.vm00:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-10T10:23:04.411 DEBUG:teuthology.orchestra.run.vm03:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-10T10:23:04.420 DEBUG:teuthology.orchestra.run.vm09:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-10T10:23:04.430 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-10T10:23:04.432 DEBUG:teuthology.orchestra.run.vm00:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-10T10:23:04.453 DEBUG:teuthology.orchestra.run.vm03:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-10T10:23:04.460 INFO:teuthology.orchestra.run.vm00.stdout:kernel.core_pattern = core 2026-03-10T10:23:04.463 DEBUG:teuthology.orchestra.run.vm09:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-10T10:23:04.469 INFO:teuthology.orchestra.run.vm03.stdout:kernel.core_pattern = core 2026-03-10T10:23:04.478 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern = core 2026-03-10T10:23:04.487 DEBUG:teuthology.orchestra.run.vm00:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-10T10:23:04.514 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:23:04.514 DEBUG:teuthology.orchestra.run.vm03:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-10T10:23:04.520 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:23:04.520 DEBUG:teuthology.orchestra.run.vm09:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-10T10:23:04.533 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:23:04.533 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-10T10:23:04.535 INFO:teuthology.task.internal:Transferring archived files... 2026-03-10T10:23:04.536 DEBUG:teuthology.misc:Transferring archived files from vm00:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/996/remote/vm00 2026-03-10T10:23:04.536 DEBUG:teuthology.orchestra.run.vm00:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-10T10:23:04.565 DEBUG:teuthology.misc:Transferring archived files from vm03:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/996/remote/vm03 2026-03-10T10:23:04.565 DEBUG:teuthology.orchestra.run.vm03:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-10T10:23:04.573 DEBUG:teuthology.misc:Transferring archived files from vm09:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/996/remote/vm09 2026-03-10T10:23:04.574 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-10T10:23:04.582 INFO:teuthology.task.internal:Removing archive directory... 2026-03-10T10:23:04.582 DEBUG:teuthology.orchestra.run.vm00:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-10T10:23:04.609 DEBUG:teuthology.orchestra.run.vm03:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-10T10:23:04.619 DEBUG:teuthology.orchestra.run.vm09:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-10T10:23:04.629 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-10T10:23:04.632 INFO:teuthology.task.internal:Not uploading archives. 2026-03-10T10:23:04.632 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-10T10:23:04.634 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-10T10:23:04.634 DEBUG:teuthology.orchestra.run.vm00:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-10T10:23:04.656 DEBUG:teuthology.orchestra.run.vm03:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-10T10:23:04.659 INFO:teuthology.orchestra.run.vm00.stdout: 258079 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 10 10:23 /home/ubuntu/cephtest 2026-03-10T10:23:04.663 DEBUG:teuthology.orchestra.run.vm09:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-10T10:23:04.665 INFO:teuthology.orchestra.run.vm03.stdout: 258078 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 10 10:23 /home/ubuntu/cephtest 2026-03-10T10:23:04.673 INFO:teuthology.orchestra.run.vm09.stdout: 258076 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 10 10:23 /home/ubuntu/cephtest 2026-03-10T10:23:04.674 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-10T10:23:04.679 INFO:teuthology.run:Summary data: description: orch/cephadm/workunits/{0-distro/ubuntu_22.04 agent/on mon_election/connectivity task/test_set_mon_crush_locations} duration: 865.242201089859 flavor: default owner: kyr success: true 2026-03-10T10:23:04.679 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-10T10:23:04.698 INFO:teuthology.run:pass